var/home/core/zuul-output/0000755000175000017500000000000015111331221014513 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015111335750015472 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000003770643615111335740017715 0ustar rootrootNov 25 13:46:29 crc systemd[1]: Starting Kubernetes Kubelet... Nov 25 13:46:29 crc restorecon[4680]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 13:46:29 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 13:46:30 crc restorecon[4680]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 13:46:30 crc restorecon[4680]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 25 13:46:31 crc kubenswrapper[4702]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 13:46:31 crc kubenswrapper[4702]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 25 13:46:31 crc kubenswrapper[4702]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 13:46:31 crc kubenswrapper[4702]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 13:46:31 crc kubenswrapper[4702]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 25 13:46:31 crc kubenswrapper[4702]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.487605 4702 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.503935 4702 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.503995 4702 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504001 4702 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504005 4702 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504011 4702 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504018 4702 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504023 4702 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504028 4702 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504032 4702 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504036 4702 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504040 4702 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504043 4702 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504046 4702 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504050 4702 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504053 4702 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504063 4702 feature_gate.go:330] unrecognized feature gate: Example Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504067 4702 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504071 4702 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504075 4702 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504079 4702 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504082 4702 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504087 4702 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504091 4702 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504095 4702 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504099 4702 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504103 4702 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504106 4702 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504110 4702 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504115 4702 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504122 4702 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504127 4702 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504132 4702 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504137 4702 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504142 4702 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504146 4702 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504150 4702 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504153 4702 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504157 4702 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504160 4702 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504164 4702 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504168 4702 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504171 4702 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504174 4702 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504178 4702 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504182 4702 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504186 4702 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504194 4702 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504198 4702 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504202 4702 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504206 4702 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504209 4702 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504213 4702 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504217 4702 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504221 4702 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504227 4702 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504272 4702 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504278 4702 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504281 4702 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504284 4702 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504288 4702 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504292 4702 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504295 4702 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504299 4702 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504302 4702 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504307 4702 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504311 4702 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504317 4702 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504321 4702 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504326 4702 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504330 4702 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.504334 4702 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504459 4702 flags.go:64] FLAG: --address="0.0.0.0" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504473 4702 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504486 4702 flags.go:64] FLAG: --anonymous-auth="true" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504492 4702 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504500 4702 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504505 4702 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504513 4702 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504520 4702 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504525 4702 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504530 4702 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504535 4702 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504541 4702 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504545 4702 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504550 4702 flags.go:64] FLAG: --cgroup-root="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504554 4702 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504559 4702 flags.go:64] FLAG: --client-ca-file="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504563 4702 flags.go:64] FLAG: --cloud-config="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504567 4702 flags.go:64] FLAG: --cloud-provider="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504571 4702 flags.go:64] FLAG: --cluster-dns="[]" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504575 4702 flags.go:64] FLAG: --cluster-domain="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504579 4702 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504584 4702 flags.go:64] FLAG: --config-dir="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504588 4702 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504592 4702 flags.go:64] FLAG: --container-log-max-files="5" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504598 4702 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504602 4702 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504607 4702 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504612 4702 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504618 4702 flags.go:64] FLAG: --contention-profiling="false" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504623 4702 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504628 4702 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504633 4702 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504638 4702 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504647 4702 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504652 4702 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504657 4702 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504662 4702 flags.go:64] FLAG: --enable-load-reader="false" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504668 4702 flags.go:64] FLAG: --enable-server="true" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504676 4702 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504689 4702 flags.go:64] FLAG: --event-burst="100" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504695 4702 flags.go:64] FLAG: --event-qps="50" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504701 4702 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504706 4702 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504711 4702 flags.go:64] FLAG: --eviction-hard="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504718 4702 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504726 4702 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504732 4702 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504738 4702 flags.go:64] FLAG: --eviction-soft="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504744 4702 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504749 4702 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504754 4702 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504759 4702 flags.go:64] FLAG: --experimental-mounter-path="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504764 4702 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504769 4702 flags.go:64] FLAG: --fail-swap-on="true" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504774 4702 flags.go:64] FLAG: --feature-gates="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504780 4702 flags.go:64] FLAG: --file-check-frequency="20s" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504786 4702 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504792 4702 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504797 4702 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504802 4702 flags.go:64] FLAG: --healthz-port="10248" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504808 4702 flags.go:64] FLAG: --help="false" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504813 4702 flags.go:64] FLAG: --hostname-override="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504818 4702 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504823 4702 flags.go:64] FLAG: --http-check-frequency="20s" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504829 4702 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504834 4702 flags.go:64] FLAG: --image-credential-provider-config="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504839 4702 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504844 4702 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504849 4702 flags.go:64] FLAG: --image-service-endpoint="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504857 4702 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504863 4702 flags.go:64] FLAG: --kube-api-burst="100" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504869 4702 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504877 4702 flags.go:64] FLAG: --kube-api-qps="50" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504883 4702 flags.go:64] FLAG: --kube-reserved="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504890 4702 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504896 4702 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504901 4702 flags.go:64] FLAG: --kubelet-cgroups="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504907 4702 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504912 4702 flags.go:64] FLAG: --lock-file="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504917 4702 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504923 4702 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504929 4702 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504949 4702 flags.go:64] FLAG: --log-json-split-stream="false" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504960 4702 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504965 4702 flags.go:64] FLAG: --log-text-split-stream="false" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504970 4702 flags.go:64] FLAG: --logging-format="text" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504974 4702 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504980 4702 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504984 4702 flags.go:64] FLAG: --manifest-url="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504988 4702 flags.go:64] FLAG: --manifest-url-header="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.504996 4702 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505001 4702 flags.go:64] FLAG: --max-open-files="1000000" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505006 4702 flags.go:64] FLAG: --max-pods="110" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505011 4702 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505016 4702 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505021 4702 flags.go:64] FLAG: --memory-manager-policy="None" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505026 4702 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505031 4702 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505035 4702 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505040 4702 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505053 4702 flags.go:64] FLAG: --node-status-max-images="50" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505057 4702 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505061 4702 flags.go:64] FLAG: --oom-score-adj="-999" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505066 4702 flags.go:64] FLAG: --pod-cidr="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505073 4702 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505081 4702 flags.go:64] FLAG: --pod-manifest-path="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505085 4702 flags.go:64] FLAG: --pod-max-pids="-1" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505090 4702 flags.go:64] FLAG: --pods-per-core="0" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505095 4702 flags.go:64] FLAG: --port="10250" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505099 4702 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505104 4702 flags.go:64] FLAG: --provider-id="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505108 4702 flags.go:64] FLAG: --qos-reserved="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505112 4702 flags.go:64] FLAG: --read-only-port="10255" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505117 4702 flags.go:64] FLAG: --register-node="true" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505121 4702 flags.go:64] FLAG: --register-schedulable="true" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505126 4702 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505136 4702 flags.go:64] FLAG: --registry-burst="10" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505140 4702 flags.go:64] FLAG: --registry-qps="5" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505144 4702 flags.go:64] FLAG: --reserved-cpus="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505150 4702 flags.go:64] FLAG: --reserved-memory="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505155 4702 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505160 4702 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505164 4702 flags.go:64] FLAG: --rotate-certificates="false" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505169 4702 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505173 4702 flags.go:64] FLAG: --runonce="false" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505177 4702 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505182 4702 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505186 4702 flags.go:64] FLAG: --seccomp-default="false" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505191 4702 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505195 4702 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505200 4702 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505205 4702 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505211 4702 flags.go:64] FLAG: --storage-driver-password="root" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505216 4702 flags.go:64] FLAG: --storage-driver-secure="false" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505222 4702 flags.go:64] FLAG: --storage-driver-table="stats" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505228 4702 flags.go:64] FLAG: --storage-driver-user="root" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505254 4702 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505260 4702 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505264 4702 flags.go:64] FLAG: --system-cgroups="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505269 4702 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505278 4702 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505282 4702 flags.go:64] FLAG: --tls-cert-file="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505286 4702 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505291 4702 flags.go:64] FLAG: --tls-min-version="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505295 4702 flags.go:64] FLAG: --tls-private-key-file="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505300 4702 flags.go:64] FLAG: --topology-manager-policy="none" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505304 4702 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505308 4702 flags.go:64] FLAG: --topology-manager-scope="container" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505313 4702 flags.go:64] FLAG: --v="2" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505320 4702 flags.go:64] FLAG: --version="false" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505328 4702 flags.go:64] FLAG: --vmodule="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505341 4702 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.505348 4702 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505469 4702 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505475 4702 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505481 4702 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505487 4702 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505494 4702 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505499 4702 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505505 4702 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505510 4702 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505515 4702 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505521 4702 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505525 4702 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505530 4702 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505535 4702 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505542 4702 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505546 4702 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505551 4702 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505556 4702 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505560 4702 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505565 4702 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505570 4702 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505576 4702 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505580 4702 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505585 4702 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505589 4702 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505593 4702 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505598 4702 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505602 4702 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505606 4702 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505610 4702 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505614 4702 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505619 4702 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505622 4702 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505626 4702 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505630 4702 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505634 4702 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505638 4702 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505643 4702 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505648 4702 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505653 4702 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505657 4702 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505660 4702 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505664 4702 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505669 4702 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505672 4702 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505676 4702 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505680 4702 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505685 4702 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505689 4702 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505693 4702 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505697 4702 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505701 4702 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505705 4702 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505709 4702 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505712 4702 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505716 4702 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505720 4702 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505723 4702 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505728 4702 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505732 4702 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505737 4702 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505742 4702 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505745 4702 feature_gate.go:330] unrecognized feature gate: Example Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505750 4702 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505754 4702 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505757 4702 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505761 4702 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505765 4702 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505768 4702 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505773 4702 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505776 4702 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.505780 4702 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.506882 4702 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.517567 4702 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.517625 4702 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517708 4702 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517717 4702 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517722 4702 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517726 4702 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517730 4702 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517733 4702 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517737 4702 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517741 4702 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517746 4702 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517755 4702 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517759 4702 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517763 4702 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517767 4702 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517771 4702 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517775 4702 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517778 4702 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517782 4702 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517786 4702 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517789 4702 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517793 4702 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517796 4702 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517801 4702 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517805 4702 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517809 4702 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517812 4702 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517816 4702 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517819 4702 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517823 4702 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517826 4702 feature_gate.go:330] unrecognized feature gate: Example Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517831 4702 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517837 4702 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517841 4702 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517845 4702 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517851 4702 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517865 4702 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517870 4702 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517874 4702 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517878 4702 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517882 4702 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517886 4702 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517890 4702 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517894 4702 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517898 4702 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517902 4702 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517906 4702 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517909 4702 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517914 4702 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517918 4702 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517922 4702 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517927 4702 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517931 4702 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517935 4702 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517939 4702 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517943 4702 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517947 4702 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517950 4702 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517954 4702 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517958 4702 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517961 4702 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517965 4702 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517969 4702 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517973 4702 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517976 4702 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517980 4702 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517984 4702 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517991 4702 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517995 4702 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.517999 4702 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518003 4702 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518019 4702 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518029 4702 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.518036 4702 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518185 4702 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518191 4702 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518196 4702 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518200 4702 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518204 4702 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518207 4702 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518211 4702 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518215 4702 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518218 4702 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518222 4702 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518225 4702 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518229 4702 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518256 4702 feature_gate.go:330] unrecognized feature gate: Example Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518260 4702 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518264 4702 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518268 4702 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518271 4702 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518275 4702 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518278 4702 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518282 4702 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518285 4702 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518290 4702 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518294 4702 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518298 4702 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518302 4702 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518306 4702 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518310 4702 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518316 4702 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518320 4702 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518324 4702 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518327 4702 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518331 4702 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518334 4702 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518338 4702 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518350 4702 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518354 4702 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518357 4702 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518361 4702 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518364 4702 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518369 4702 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518374 4702 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518379 4702 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518383 4702 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518386 4702 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518390 4702 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518394 4702 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518397 4702 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518402 4702 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518407 4702 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518410 4702 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518414 4702 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518418 4702 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518422 4702 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518426 4702 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518430 4702 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518434 4702 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518438 4702 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518442 4702 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518445 4702 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518449 4702 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518452 4702 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518456 4702 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518459 4702 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518463 4702 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518466 4702 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518470 4702 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518473 4702 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518477 4702 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518480 4702 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518483 4702 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.518494 4702 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.518501 4702 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.518731 4702 server.go:940] "Client rotation is on, will bootstrap in background" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.533812 4702 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.533955 4702 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.536490 4702 server.go:997] "Starting client certificate rotation" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.536520 4702 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.539293 4702 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-30 01:35:15.955035684 +0000 UTC Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.539398 4702 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 827h48m44.415641113s for next certificate rotation Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.598348 4702 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.600528 4702 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.629457 4702 log.go:25] "Validated CRI v1 runtime API" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.685502 4702 log.go:25] "Validated CRI v1 image API" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.687673 4702 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.700618 4702 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-25-13-42-50-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.700659 4702 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:41 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.717071 4702 manager.go:217] Machine: {Timestamp:2025-11-25 13:46:31.714896131 +0000 UTC m=+0.877546611 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:bb366f18-cc3c-451f-9d6a-c0cc41864b99 BootID:96cc93a3-66f3-477f-bbc8-ef3a984ed406 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:41 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:85:e9:60 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:85:e9:60 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:10:9d:c9 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:f2:e1:15 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:12:8f:23 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:af:a4:43 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:fa:ca:a6:1c:ff:f0 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:8a:01:95:45:62:39 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.717398 4702 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.717574 4702 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.717924 4702 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.718265 4702 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.718305 4702 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.719276 4702 topology_manager.go:138] "Creating topology manager with none policy" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.719299 4702 container_manager_linux.go:303] "Creating device plugin manager" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.719999 4702 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.720030 4702 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.720826 4702 state_mem.go:36] "Initialized new in-memory state store" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.721396 4702 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.724918 4702 kubelet.go:418] "Attempting to sync node with API server" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.724951 4702 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.724992 4702 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.725010 4702 kubelet.go:324] "Adding apiserver pod source" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.725025 4702 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.732062 4702 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.732068 4702 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Nov 25 13:46:31 crc kubenswrapper[4702]: E1125 13:46:31.732223 4702 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Nov 25 13:46:31 crc kubenswrapper[4702]: E1125 13:46:31.732319 4702 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.732838 4702 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.733802 4702 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.735065 4702 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.736316 4702 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.736342 4702 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.736352 4702 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.736361 4702 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.736374 4702 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.736383 4702 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.736391 4702 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.736403 4702 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.736412 4702 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.736421 4702 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.736447 4702 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.736455 4702 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.737596 4702 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.738180 4702 server.go:1280] "Started kubelet" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.738321 4702 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.738415 4702 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.738413 4702 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.739401 4702 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 25 13:46:31 crc systemd[1]: Started Kubernetes Kubelet. Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.743378 4702 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.743427 4702 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.743677 4702 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 10:57:10.114534894 +0000 UTC Nov 25 13:46:31 crc kubenswrapper[4702]: E1125 13:46:31.743788 4702 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.743790 4702 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 525h10m38.370778233s for next certificate rotation Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.744800 4702 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.744824 4702 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.744894 4702 server.go:460] "Adding debug handlers to kubelet server" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.744899 4702 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.745091 4702 factory.go:153] Registering CRI-O factory Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.745122 4702 factory.go:221] Registration of the crio container factory successfully Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.745279 4702 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.745315 4702 factory.go:55] Registering systemd factory Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.745326 4702 factory.go:221] Registration of the systemd container factory successfully Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.745359 4702 factory.go:103] Registering Raw factory Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.745379 4702 manager.go:1196] Started watching for new ooms in manager Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.745362 4702 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Nov 25 13:46:31 crc kubenswrapper[4702]: E1125 13:46:31.745418 4702 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Nov 25 13:46:31 crc kubenswrapper[4702]: E1125 13:46:31.747955 4702 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="200ms" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.752904 4702 manager.go:319] Starting recovery of all containers Nov 25 13:46:31 crc kubenswrapper[4702]: E1125 13:46:31.752987 4702 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.180:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187b43f3677eaf48 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 13:46:31.738150728 +0000 UTC m=+0.900801178,LastTimestamp:2025-11-25 13:46:31.738150728 +0000 UTC m=+0.900801178,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.755771 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.755844 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.755920 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.755936 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.755949 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.755960 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.755971 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.756020 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.756040 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.756059 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.756068 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.756088 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.756097 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.756109 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.756118 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.756131 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.756140 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.756149 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.756160 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.756168 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.756177 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.756187 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.756199 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.756251 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.756261 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.756271 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.756286 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.756297 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.756307 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.756317 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.756327 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.757175 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.757777 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.757805 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.757820 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.757836 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.757849 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.757863 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.757877 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.757892 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.757905 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.757917 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.757934 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.757960 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.757973 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.757986 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758000 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758017 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758032 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758046 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758060 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758072 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758091 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758108 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758123 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758138 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758151 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758166 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758180 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758192 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758205 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758217 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758248 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758268 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758281 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758296 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758308 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758321 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758335 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758347 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758360 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758374 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758388 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758400 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758413 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758425 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758438 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758454 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758469 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758483 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758496 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758508 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758521 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758534 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.758549 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.761046 4702 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.761121 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.762258 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.762296 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.762320 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.762336 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.762354 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.762371 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.762389 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.762404 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.762423 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.762438 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.762453 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.763029 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.763210 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.763386 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.763424 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.763440 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.764389 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.765011 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.765186 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.765213 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.765253 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.765281 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.765299 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.765452 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.765525 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.765541 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.765554 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.765569 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.765585 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.765608 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.765622 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.765636 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.765684 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.765703 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.765722 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.765741 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.765759 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.765775 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.765817 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.765838 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.765876 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.765890 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.765904 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.765970 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.765984 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.765997 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766016 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766026 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766054 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766063 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766074 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766115 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766125 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766184 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766271 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766299 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766336 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766347 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766382 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766395 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766406 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766418 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766431 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766448 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766504 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766555 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766569 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766582 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766595 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766609 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766647 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766662 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766689 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766708 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766722 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766735 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766746 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766757 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766769 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766812 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766849 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766863 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766876 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766890 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766906 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766925 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766936 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.766948 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.767201 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.767222 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.767277 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.767290 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.767386 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.767401 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.767414 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.767425 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.767456 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.767469 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.767483 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.767515 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.767527 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.767538 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.768659 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.768768 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.768893 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.768994 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.769078 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.769197 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.769417 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.769509 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.769624 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.769878 4702 manager.go:324] Recovery completed Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.770700 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.770790 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.770852 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.770916 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.770975 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.771090 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.771160 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.771224 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.771353 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.771504 4702 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.771600 4702 reconstruct.go:97] "Volume reconstruction finished" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.772260 4702 reconciler.go:26] "Reconciler: start to sync state" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.784920 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.787018 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.787060 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.787069 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.788110 4702 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.788132 4702 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.788155 4702 state_mem.go:36] "Initialized new in-memory state store" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.797826 4702 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.799729 4702 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.799765 4702 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.799788 4702 kubelet.go:2335] "Starting kubelet main sync loop" Nov 25 13:46:31 crc kubenswrapper[4702]: E1125 13:46:31.799847 4702 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 25 13:46:31 crc kubenswrapper[4702]: W1125 13:46:31.800774 4702 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Nov 25 13:46:31 crc kubenswrapper[4702]: E1125 13:46:31.802337 4702 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.803674 4702 policy_none.go:49] "None policy: Start" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.805274 4702 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.805353 4702 state_mem.go:35] "Initializing new in-memory state store" Nov 25 13:46:31 crc kubenswrapper[4702]: E1125 13:46:31.844321 4702 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 25 13:46:31 crc kubenswrapper[4702]: E1125 13:46:31.900349 4702 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Nov 25 13:46:31 crc kubenswrapper[4702]: E1125 13:46:31.944595 4702 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 25 13:46:31 crc kubenswrapper[4702]: E1125 13:46:31.949262 4702 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="400ms" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.989480 4702 manager.go:334] "Starting Device Plugin manager" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.989558 4702 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.989581 4702 server.go:79] "Starting device plugin registration server" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.990120 4702 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.990152 4702 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.990443 4702 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.990590 4702 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 25 13:46:31 crc kubenswrapper[4702]: I1125 13:46:31.990605 4702 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 25 13:46:32 crc kubenswrapper[4702]: E1125 13:46:32.001289 4702 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.091126 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.092381 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.092468 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.092483 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.092532 4702 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 13:46:32 crc kubenswrapper[4702]: E1125 13:46:32.093338 4702 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.180:6443: connect: connection refused" node="crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.101161 4702 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.101263 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.102359 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.102395 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.102405 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.102567 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.102876 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.102935 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.103306 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.103345 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.103354 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.103463 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.103740 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.103775 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.104005 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.104024 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.104032 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.104350 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.104375 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.104387 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.104431 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.104445 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.104454 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.104632 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.104811 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.104856 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.105287 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.105312 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.105323 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.105421 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.105599 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.105641 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.105867 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.105889 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.105900 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.106160 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.106185 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.106196 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.106401 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.106439 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.106475 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.106494 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.106504 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.107203 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.107253 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.107277 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.177377 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.177425 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.177451 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.177470 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.177535 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.177596 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.177634 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.177661 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.177683 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.177745 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.177779 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.177807 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.177831 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.177869 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.177891 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.278684 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.278755 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.278914 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.278977 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.278954 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.279075 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.279099 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.279103 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.278903 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.279121 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.279141 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.279147 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.279180 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.279196 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.279212 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.279296 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.279319 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.279334 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.279351 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.279392 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.279413 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.279434 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.279440 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.279485 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.279486 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.279506 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.279543 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.279548 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.279563 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.279595 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.294310 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.295892 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.295942 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.295951 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.295981 4702 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 13:46:32 crc kubenswrapper[4702]: E1125 13:46:32.296634 4702 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.180:6443: connect: connection refused" node="crc" Nov 25 13:46:32 crc kubenswrapper[4702]: E1125 13:46:32.350643 4702 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="800ms" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.439627 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.446971 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.471280 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.492124 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.497434 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 13:46:32 crc kubenswrapper[4702]: W1125 13:46:32.597618 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-78679c797a082ecec1fbf4711878db898a520f0d0d1816bf41a47e0e1dc8d08c WatchSource:0}: Error finding container 78679c797a082ecec1fbf4711878db898a520f0d0d1816bf41a47e0e1dc8d08c: Status 404 returned error can't find the container with id 78679c797a082ecec1fbf4711878db898a520f0d0d1816bf41a47e0e1dc8d08c Nov 25 13:46:32 crc kubenswrapper[4702]: W1125 13:46:32.604400 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-10e0d879ba58d660760ec03a32838245aa9b93cb163188ddcc172eb220d78381 WatchSource:0}: Error finding container 10e0d879ba58d660760ec03a32838245aa9b93cb163188ddcc172eb220d78381: Status 404 returned error can't find the container with id 10e0d879ba58d660760ec03a32838245aa9b93cb163188ddcc172eb220d78381 Nov 25 13:46:32 crc kubenswrapper[4702]: W1125 13:46:32.605035 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-fd47ac2b1c7877128d2cdb6b0d399186a2aed81431d627f22657f466b0e4904d WatchSource:0}: Error finding container fd47ac2b1c7877128d2cdb6b0d399186a2aed81431d627f22657f466b0e4904d: Status 404 returned error can't find the container with id fd47ac2b1c7877128d2cdb6b0d399186a2aed81431d627f22657f466b0e4904d Nov 25 13:46:32 crc kubenswrapper[4702]: W1125 13:46:32.606286 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-6e7fdd9973449e7122bdf359cd19ebf6514e96915b2bd7825ec326e0a6f9c4f4 WatchSource:0}: Error finding container 6e7fdd9973449e7122bdf359cd19ebf6514e96915b2bd7825ec326e0a6f9c4f4: Status 404 returned error can't find the container with id 6e7fdd9973449e7122bdf359cd19ebf6514e96915b2bd7825ec326e0a6f9c4f4 Nov 25 13:46:32 crc kubenswrapper[4702]: W1125 13:46:32.653716 4702 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Nov 25 13:46:32 crc kubenswrapper[4702]: E1125 13:46:32.653837 4702 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.697967 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.699551 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.699608 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.699622 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.699654 4702 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 13:46:32 crc kubenswrapper[4702]: E1125 13:46:32.700372 4702 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.180:6443: connect: connection refused" node="crc" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.739966 4702 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Nov 25 13:46:32 crc kubenswrapper[4702]: W1125 13:46:32.766997 4702 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Nov 25 13:46:32 crc kubenswrapper[4702]: E1125 13:46:32.767098 4702 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.804904 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"89814094dc13daaa275010739ba2fb390b866b75149f8ef4a790ffdc3c76a90a"} Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.805708 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"78679c797a082ecec1fbf4711878db898a520f0d0d1816bf41a47e0e1dc8d08c"} Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.806652 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"fd47ac2b1c7877128d2cdb6b0d399186a2aed81431d627f22657f466b0e4904d"} Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.807910 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"6e7fdd9973449e7122bdf359cd19ebf6514e96915b2bd7825ec326e0a6f9c4f4"} Nov 25 13:46:32 crc kubenswrapper[4702]: I1125 13:46:32.808713 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"10e0d879ba58d660760ec03a32838245aa9b93cb163188ddcc172eb220d78381"} Nov 25 13:46:32 crc kubenswrapper[4702]: W1125 13:46:32.902433 4702 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Nov 25 13:46:32 crc kubenswrapper[4702]: E1125 13:46:32.902521 4702 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Nov 25 13:46:33 crc kubenswrapper[4702]: E1125 13:46:33.152558 4702 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="1.6s" Nov 25 13:46:33 crc kubenswrapper[4702]: W1125 13:46:33.162223 4702 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Nov 25 13:46:33 crc kubenswrapper[4702]: E1125 13:46:33.162337 4702 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.500504 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.502212 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.502285 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.502300 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.502338 4702 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 13:46:33 crc kubenswrapper[4702]: E1125 13:46:33.503092 4702 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.180:6443: connect: connection refused" node="crc" Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.740460 4702 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.813216 4702 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="5aaf34113a5a5c1948b672848f8bbcbc0ef69e6a3efe03202c9210016a0c6980" exitCode=0 Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.813357 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.813396 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"5aaf34113a5a5c1948b672848f8bbcbc0ef69e6a3efe03202c9210016a0c6980"} Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.815069 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.815110 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.815122 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.815581 4702 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92" exitCode=0 Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.815678 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92"} Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.815766 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.816771 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.816799 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.816811 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.817952 4702 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512" exitCode=0 Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.818034 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.818036 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512"} Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.818356 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.819261 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.819300 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.819313 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.819582 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.819611 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.819621 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.820159 4702 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="e08782f884652e08e363cd5cad0db72b87dedb7eaefa8db537a43ebdfc778325" exitCode=0 Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.820247 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"e08782f884652e08e363cd5cad0db72b87dedb7eaefa8db537a43ebdfc778325"} Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.820283 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.821082 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.821111 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.821124 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:33 crc kubenswrapper[4702]: I1125 13:46:33.821865 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773"} Nov 25 13:46:34 crc kubenswrapper[4702]: W1125 13:46:34.439539 4702 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Nov 25 13:46:34 crc kubenswrapper[4702]: E1125 13:46:34.439642 4702 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Nov 25 13:46:34 crc kubenswrapper[4702]: I1125 13:46:34.739757 4702 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Nov 25 13:46:34 crc kubenswrapper[4702]: E1125 13:46:34.753833 4702 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="3.2s" Nov 25 13:46:34 crc kubenswrapper[4702]: I1125 13:46:34.831595 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b961d759f8b153e4c876e6f1b5eff499fc72a3c828610fc980cba592cac76015"} Nov 25 13:46:34 crc kubenswrapper[4702]: I1125 13:46:34.831674 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b01553a55ccc3121476a5fdfed5532047ecd1e46545aab873da4a7d4718bd96f"} Nov 25 13:46:34 crc kubenswrapper[4702]: I1125 13:46:34.831687 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ff40c37e9756ca143489aafc21e756e213d7204f041b8b807eef932de448210a"} Nov 25 13:46:34 crc kubenswrapper[4702]: I1125 13:46:34.831745 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:34 crc kubenswrapper[4702]: I1125 13:46:34.832982 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:34 crc kubenswrapper[4702]: I1125 13:46:34.833016 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:34 crc kubenswrapper[4702]: I1125 13:46:34.833030 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:34 crc kubenswrapper[4702]: I1125 13:46:34.834962 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0"} Nov 25 13:46:34 crc kubenswrapper[4702]: I1125 13:46:34.834998 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b"} Nov 25 13:46:34 crc kubenswrapper[4702]: I1125 13:46:34.835009 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057"} Nov 25 13:46:34 crc kubenswrapper[4702]: I1125 13:46:34.835027 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec"} Nov 25 13:46:34 crc kubenswrapper[4702]: I1125 13:46:34.837032 4702 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158" exitCode=0 Nov 25 13:46:34 crc kubenswrapper[4702]: I1125 13:46:34.837124 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158"} Nov 25 13:46:34 crc kubenswrapper[4702]: I1125 13:46:34.837136 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:34 crc kubenswrapper[4702]: I1125 13:46:34.838100 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:34 crc kubenswrapper[4702]: I1125 13:46:34.838129 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:34 crc kubenswrapper[4702]: I1125 13:46:34.838142 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:34 crc kubenswrapper[4702]: I1125 13:46:34.839352 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"90db72f00d2ebd6a338e6ec5396349087083c2ad55dd120e712b1e4916b7c5e0"} Nov 25 13:46:34 crc kubenswrapper[4702]: I1125 13:46:34.839432 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:34 crc kubenswrapper[4702]: I1125 13:46:34.840332 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:34 crc kubenswrapper[4702]: I1125 13:46:34.840349 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:34 crc kubenswrapper[4702]: I1125 13:46:34.840357 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:34 crc kubenswrapper[4702]: I1125 13:46:34.844180 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2"} Nov 25 13:46:34 crc kubenswrapper[4702]: I1125 13:46:34.844207 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658"} Nov 25 13:46:34 crc kubenswrapper[4702]: I1125 13:46:34.844218 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a"} Nov 25 13:46:34 crc kubenswrapper[4702]: I1125 13:46:34.844313 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:34 crc kubenswrapper[4702]: I1125 13:46:34.847633 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:34 crc kubenswrapper[4702]: I1125 13:46:34.847659 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:34 crc kubenswrapper[4702]: I1125 13:46:34.847667 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:35 crc kubenswrapper[4702]: I1125 13:46:35.103560 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:35 crc kubenswrapper[4702]: I1125 13:46:35.105743 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:35 crc kubenswrapper[4702]: I1125 13:46:35.105812 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:35 crc kubenswrapper[4702]: I1125 13:46:35.105827 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:35 crc kubenswrapper[4702]: I1125 13:46:35.105862 4702 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 13:46:35 crc kubenswrapper[4702]: E1125 13:46:35.106522 4702 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.180:6443: connect: connection refused" node="crc" Nov 25 13:46:35 crc kubenswrapper[4702]: W1125 13:46:35.314068 4702 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Nov 25 13:46:35 crc kubenswrapper[4702]: E1125 13:46:35.314183 4702 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Nov 25 13:46:35 crc kubenswrapper[4702]: W1125 13:46:35.330401 4702 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Nov 25 13:46:35 crc kubenswrapper[4702]: E1125 13:46:35.330517 4702 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Nov 25 13:46:35 crc kubenswrapper[4702]: I1125 13:46:35.849496 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4"} Nov 25 13:46:35 crc kubenswrapper[4702]: I1125 13:46:35.849641 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:35 crc kubenswrapper[4702]: I1125 13:46:35.854033 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:35 crc kubenswrapper[4702]: I1125 13:46:35.854069 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:35 crc kubenswrapper[4702]: I1125 13:46:35.854080 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:35 crc kubenswrapper[4702]: I1125 13:46:35.855816 4702 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e" exitCode=0 Nov 25 13:46:35 crc kubenswrapper[4702]: I1125 13:46:35.855841 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e"} Nov 25 13:46:35 crc kubenswrapper[4702]: I1125 13:46:35.855881 4702 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 13:46:35 crc kubenswrapper[4702]: I1125 13:46:35.855905 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:35 crc kubenswrapper[4702]: I1125 13:46:35.855916 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:35 crc kubenswrapper[4702]: I1125 13:46:35.855920 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:35 crc kubenswrapper[4702]: I1125 13:46:35.856015 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:35 crc kubenswrapper[4702]: I1125 13:46:35.856769 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:35 crc kubenswrapper[4702]: I1125 13:46:35.856797 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:35 crc kubenswrapper[4702]: I1125 13:46:35.856804 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:35 crc kubenswrapper[4702]: I1125 13:46:35.856959 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:35 crc kubenswrapper[4702]: I1125 13:46:35.857015 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:35 crc kubenswrapper[4702]: I1125 13:46:35.857033 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:35 crc kubenswrapper[4702]: I1125 13:46:35.857342 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:35 crc kubenswrapper[4702]: I1125 13:46:35.857375 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:35 crc kubenswrapper[4702]: I1125 13:46:35.857387 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:35 crc kubenswrapper[4702]: I1125 13:46:35.857438 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:35 crc kubenswrapper[4702]: I1125 13:46:35.857451 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:35 crc kubenswrapper[4702]: I1125 13:46:35.857459 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:36 crc kubenswrapper[4702]: I1125 13:46:36.564777 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 13:46:36 crc kubenswrapper[4702]: I1125 13:46:36.575490 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:46:36 crc kubenswrapper[4702]: I1125 13:46:36.865932 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d"} Nov 25 13:46:36 crc kubenswrapper[4702]: I1125 13:46:36.866047 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d"} Nov 25 13:46:36 crc kubenswrapper[4702]: I1125 13:46:36.866064 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1"} Nov 25 13:46:36 crc kubenswrapper[4702]: I1125 13:46:36.866067 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:36 crc kubenswrapper[4702]: I1125 13:46:36.866087 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:36 crc kubenswrapper[4702]: I1125 13:46:36.866214 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:36 crc kubenswrapper[4702]: I1125 13:46:36.866077 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926"} Nov 25 13:46:36 crc kubenswrapper[4702]: I1125 13:46:36.866518 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376"} Nov 25 13:46:36 crc kubenswrapper[4702]: I1125 13:46:36.866579 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:46:36 crc kubenswrapper[4702]: I1125 13:46:36.867580 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:36 crc kubenswrapper[4702]: I1125 13:46:36.867622 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:36 crc kubenswrapper[4702]: I1125 13:46:36.867582 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:36 crc kubenswrapper[4702]: I1125 13:46:36.867656 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:36 crc kubenswrapper[4702]: I1125 13:46:36.867665 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:36 crc kubenswrapper[4702]: I1125 13:46:36.867633 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:36 crc kubenswrapper[4702]: I1125 13:46:36.867913 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:36 crc kubenswrapper[4702]: I1125 13:46:36.867970 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:36 crc kubenswrapper[4702]: I1125 13:46:36.867981 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:37 crc kubenswrapper[4702]: I1125 13:46:37.051589 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 13:46:37 crc kubenswrapper[4702]: I1125 13:46:37.051785 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:37 crc kubenswrapper[4702]: I1125 13:46:37.053174 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:37 crc kubenswrapper[4702]: I1125 13:46:37.053207 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:37 crc kubenswrapper[4702]: I1125 13:46:37.053215 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:37 crc kubenswrapper[4702]: I1125 13:46:37.868006 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:37 crc kubenswrapper[4702]: I1125 13:46:37.868058 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:37 crc kubenswrapper[4702]: I1125 13:46:37.869673 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:37 crc kubenswrapper[4702]: I1125 13:46:37.869711 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:37 crc kubenswrapper[4702]: I1125 13:46:37.869673 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:37 crc kubenswrapper[4702]: I1125 13:46:37.869724 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:37 crc kubenswrapper[4702]: I1125 13:46:37.869743 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:37 crc kubenswrapper[4702]: I1125 13:46:37.869756 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:38 crc kubenswrapper[4702]: I1125 13:46:38.307640 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:38 crc kubenswrapper[4702]: I1125 13:46:38.309297 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:38 crc kubenswrapper[4702]: I1125 13:46:38.309340 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:38 crc kubenswrapper[4702]: I1125 13:46:38.309349 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:38 crc kubenswrapper[4702]: I1125 13:46:38.309371 4702 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 13:46:39 crc kubenswrapper[4702]: I1125 13:46:39.208016 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 25 13:46:39 crc kubenswrapper[4702]: I1125 13:46:39.208277 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:39 crc kubenswrapper[4702]: I1125 13:46:39.209543 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:39 crc kubenswrapper[4702]: I1125 13:46:39.209600 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:39 crc kubenswrapper[4702]: I1125 13:46:39.209620 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:39 crc kubenswrapper[4702]: I1125 13:46:39.565581 4702 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 25 13:46:39 crc kubenswrapper[4702]: I1125 13:46:39.565695 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 13:46:40 crc kubenswrapper[4702]: I1125 13:46:40.188974 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:46:40 crc kubenswrapper[4702]: I1125 13:46:40.189320 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:40 crc kubenswrapper[4702]: I1125 13:46:40.191190 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:40 crc kubenswrapper[4702]: I1125 13:46:40.191277 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:40 crc kubenswrapper[4702]: I1125 13:46:40.191293 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:40 crc kubenswrapper[4702]: I1125 13:46:40.307971 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 13:46:40 crc kubenswrapper[4702]: I1125 13:46:40.308444 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:40 crc kubenswrapper[4702]: I1125 13:46:40.310061 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:40 crc kubenswrapper[4702]: I1125 13:46:40.310096 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:40 crc kubenswrapper[4702]: I1125 13:46:40.310106 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:40 crc kubenswrapper[4702]: I1125 13:46:40.318837 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 13:46:40 crc kubenswrapper[4702]: I1125 13:46:40.875278 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:40 crc kubenswrapper[4702]: I1125 13:46:40.875395 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 13:46:40 crc kubenswrapper[4702]: I1125 13:46:40.876271 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:40 crc kubenswrapper[4702]: I1125 13:46:40.876309 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:40 crc kubenswrapper[4702]: I1125 13:46:40.876322 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:41 crc kubenswrapper[4702]: I1125 13:46:41.878199 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:41 crc kubenswrapper[4702]: I1125 13:46:41.879454 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:41 crc kubenswrapper[4702]: I1125 13:46:41.879525 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:41 crc kubenswrapper[4702]: I1125 13:46:41.879547 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:42 crc kubenswrapper[4702]: E1125 13:46:42.001448 4702 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 25 13:46:43 crc kubenswrapper[4702]: I1125 13:46:43.086578 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 13:46:43 crc kubenswrapper[4702]: I1125 13:46:43.086753 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:43 crc kubenswrapper[4702]: I1125 13:46:43.088061 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:43 crc kubenswrapper[4702]: I1125 13:46:43.088138 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:43 crc kubenswrapper[4702]: I1125 13:46:43.088158 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:44 crc kubenswrapper[4702]: I1125 13:46:44.666917 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 13:46:44 crc kubenswrapper[4702]: I1125 13:46:44.667135 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:44 crc kubenswrapper[4702]: I1125 13:46:44.668723 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:44 crc kubenswrapper[4702]: I1125 13:46:44.668791 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:44 crc kubenswrapper[4702]: I1125 13:46:44.668805 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:44 crc kubenswrapper[4702]: I1125 13:46:44.892362 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 25 13:46:44 crc kubenswrapper[4702]: I1125 13:46:44.892614 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:44 crc kubenswrapper[4702]: I1125 13:46:44.893976 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:44 crc kubenswrapper[4702]: I1125 13:46:44.894024 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:44 crc kubenswrapper[4702]: I1125 13:46:44.894036 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:45 crc kubenswrapper[4702]: I1125 13:46:45.480671 4702 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Nov 25 13:46:45 crc kubenswrapper[4702]: I1125 13:46:45.480747 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 25 13:46:45 crc kubenswrapper[4702]: I1125 13:46:45.486369 4702 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Nov 25 13:46:45 crc kubenswrapper[4702]: I1125 13:46:45.486422 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 25 13:46:49 crc kubenswrapper[4702]: I1125 13:46:49.565917 4702 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 25 13:46:49 crc kubenswrapper[4702]: I1125 13:46:49.566036 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.194779 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.194981 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.196138 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.196169 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.196185 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.199229 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:46:50 crc kubenswrapper[4702]: E1125 13:46:50.484534 4702 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.486059 4702 trace.go:236] Trace[263506295]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 13:46:35.639) (total time: 14846ms): Nov 25 13:46:50 crc kubenswrapper[4702]: Trace[263506295]: ---"Objects listed" error: 14846ms (13:46:50.485) Nov 25 13:46:50 crc kubenswrapper[4702]: Trace[263506295]: [14.846182664s] [14.846182664s] END Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.486081 4702 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.486693 4702 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.490414 4702 trace.go:236] Trace[326263502]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 13:46:39.811) (total time: 10679ms): Nov 25 13:46:50 crc kubenswrapper[4702]: Trace[326263502]: ---"Objects listed" error: 10678ms (13:46:50.490) Nov 25 13:46:50 crc kubenswrapper[4702]: Trace[326263502]: [10.679055s] [10.679055s] END Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.490462 4702 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 25 13:46:50 crc kubenswrapper[4702]: E1125 13:46:50.490538 4702 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.492012 4702 trace.go:236] Trace[1283614171]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 13:46:39.719) (total time: 10772ms): Nov 25 13:46:50 crc kubenswrapper[4702]: Trace[1283614171]: ---"Objects listed" error: 10771ms (13:46:50.491) Nov 25 13:46:50 crc kubenswrapper[4702]: Trace[1283614171]: [10.772158132s] [10.772158132s] END Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.492431 4702 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.528680 4702 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.535004 4702 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": EOF" start-of-body= Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.535048 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": EOF" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.535439 4702 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": EOF" start-of-body= Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.535663 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": EOF" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.536710 4702 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": write tcp 192.168.126.11:47426->192.168.126.11:17697: write: connection reset by peer" start-of-body= Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.536800 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": write tcp 192.168.126.11:47426->192.168.126.11:17697: write: connection reset by peer" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.736132 4702 apiserver.go:52] "Watching apiserver" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.739632 4702 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.739914 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.740265 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.740415 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.740477 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.740613 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.740849 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 13:46:50 crc kubenswrapper[4702]: E1125 13:46:50.741048 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.740893 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:46:50 crc kubenswrapper[4702]: E1125 13:46:50.741093 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:46:50 crc kubenswrapper[4702]: E1125 13:46:50.741159 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.742068 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.742652 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.743055 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.743129 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.743180 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.743198 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.743288 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.743336 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.743460 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.746309 4702 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.772785 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.780928 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.790887 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.800953 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.808828 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.817546 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.826922 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.830346 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.830404 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.830433 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.830457 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.830484 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.830506 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.830528 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.830554 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.830580 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.830603 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.830630 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.830652 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.830700 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.830715 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.830737 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.830761 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.830784 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.830808 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.830828 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.830851 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.830874 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.830903 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.830860 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.830913 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831017 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831044 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831047 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831092 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831113 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831133 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831152 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831174 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831192 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831210 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831256 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831275 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831293 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831335 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831355 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831371 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831390 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831409 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831432 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831452 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831493 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831509 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831525 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831541 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831559 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831576 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831592 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831608 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831628 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831667 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831683 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831700 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831716 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831735 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831752 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831771 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832020 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832058 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832073 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832090 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832107 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832122 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832137 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832155 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832172 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832187 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832203 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832218 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832247 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832276 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832303 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832324 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832343 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832359 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832378 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832417 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832436 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.834251 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.834284 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.834326 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.834352 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.834378 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.834401 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.834422 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.834449 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.834485 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.834507 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.834529 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.834550 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.834574 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.834593 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.834615 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.834648 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.834674 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.834698 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.834809 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.834836 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.834875 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.834900 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.834923 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.834945 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.834969 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.835004 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.835028 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.835054 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.835078 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.835122 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.835704 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.835955 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.835986 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.836145 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.836172 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.836202 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.836765 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.837001 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.837165 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.837262 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.837300 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.837325 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.837351 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.837375 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.837399 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.837422 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.837449 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.837476 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.837500 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.837527 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.837551 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.837585 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.837612 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.837637 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.837660 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.837690 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.837713 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.837934 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.837962 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.837987 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.838011 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.838033 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.838055 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.838082 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.838108 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.838180 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.838206 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.838227 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831427 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.842355 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831610 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831706 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831755 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.831962 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832072 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832126 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832304 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832376 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832474 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832438 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832681 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832708 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832749 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832789 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832807 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832896 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832918 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.832955 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.833201 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.833255 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.833276 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.842545 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.833354 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.833492 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.833602 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.833618 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.842580 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.833676 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.833677 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.833778 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.833957 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.833977 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.834040 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.834119 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.834178 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.834302 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.834427 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.834641 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.834737 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.834869 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.835048 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.835209 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.835570 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.835597 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.842820 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.835615 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.835718 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.836071 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.836108 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.836494 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.836610 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.836630 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.836634 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.836654 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.836379 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.836776 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.837392 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.837453 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.837508 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.837515 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.837727 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.838069 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.838123 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.838282 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.838417 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: E1125 13:46:50.838489 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:46:51.338463408 +0000 UTC m=+20.501113858 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.838724 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.839310 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.839501 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.839513 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.839627 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.839636 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.839641 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.839891 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.839962 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.840030 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.840036 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.840218 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.840311 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.840322 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.840335 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.840458 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.840542 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.840609 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.840640 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.840870 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.841035 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.840878 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.841162 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.841192 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.841334 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.841351 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.841460 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.841438 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.840975 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.841653 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.841983 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.842183 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.842263 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.842260 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.842458 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.842558 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.842576 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.833701 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.842885 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843308 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843348 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843380 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843412 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843443 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843468 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843493 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843520 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843545 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843572 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843597 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843620 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843643 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843667 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843693 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843717 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843746 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843774 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843797 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843820 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843845 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843869 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843896 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843920 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843943 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843966 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843988 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.844012 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.844035 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.844056 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.844079 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.844105 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.844131 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.844154 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.844176 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.844200 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.844222 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.844276 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.844300 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.844322 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.844350 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.844380 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843346 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.844717 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843359 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843583 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843594 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843683 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843737 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.843838 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.844151 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.844207 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.844379 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.844410 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.844603 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.844699 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.844693 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.844911 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.845200 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.845301 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.845302 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.845523 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.845781 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.845969 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.846034 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.845979 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.846058 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.846070 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.846101 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.846129 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.846156 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.846185 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.846210 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.846260 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.846293 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.846361 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.846403 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.846441 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.846473 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.846502 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.846543 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.846578 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.846474 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.846606 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.846640 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.846680 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.846706 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.846722 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.846734 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847044 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847127 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847162 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847266 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847282 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847302 4702 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847312 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847324 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847332 4702 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847343 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847353 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847363 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847372 4702 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847381 4702 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847390 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847400 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847409 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847418 4702 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847430 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847440 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847449 4702 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847457 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847467 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847475 4702 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847486 4702 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847497 4702 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847505 4702 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847514 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847522 4702 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847531 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847540 4702 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847550 4702 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847558 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847567 4702 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847578 4702 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847587 4702 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847596 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847606 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847613 4702 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847621 4702 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: E1125 13:46:50.847286 4702 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847648 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.847754 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: E1125 13:46:50.848028 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 13:46:51.348003887 +0000 UTC m=+20.510654337 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.848115 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.848127 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: E1125 13:46:50.848306 4702 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.848331 4702 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.848369 4702 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.848398 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: E1125 13:46:50.848475 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 13:46:51.348448589 +0000 UTC m=+20.511099039 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.848819 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.848851 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.849053 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.849079 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850002 4702 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850025 4702 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850035 4702 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850071 4702 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850082 4702 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850092 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850122 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850136 4702 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850168 4702 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850189 4702 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850202 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850254 4702 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850270 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850278 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850288 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850300 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850311 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850322 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850333 4702 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850343 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850353 4702 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850362 4702 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850374 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850384 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850395 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850403 4702 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850413 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850421 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850430 4702 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850440 4702 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850451 4702 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850459 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850469 4702 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850479 4702 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850489 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850498 4702 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850506 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850516 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850526 4702 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850534 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850542 4702 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850551 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850560 4702 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850571 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850592 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850605 4702 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850618 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850631 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850647 4702 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850662 4702 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850678 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850688 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850698 4702 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850708 4702 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850718 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850727 4702 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850737 4702 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850746 4702 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850756 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850765 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850775 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850784 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850793 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850803 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850814 4702 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850824 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850834 4702 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850843 4702 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850852 4702 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850861 4702 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850870 4702 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850880 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850890 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850898 4702 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850907 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850917 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850927 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850936 4702 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850945 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850955 4702 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850963 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850973 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850982 4702 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.850992 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.851002 4702 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.851011 4702 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.851021 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.851030 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.851039 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.851049 4702 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.851058 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.851067 4702 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.851076 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.851086 4702 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.851094 4702 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.851103 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.851111 4702 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.854741 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: E1125 13:46:50.861756 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 13:46:50 crc kubenswrapper[4702]: E1125 13:46:50.861798 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 13:46:50 crc kubenswrapper[4702]: E1125 13:46:50.861820 4702 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 13:46:50 crc kubenswrapper[4702]: E1125 13:46:50.861906 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 13:46:51.361880854 +0000 UTC m=+20.524531294 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.864518 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.864627 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: E1125 13:46:50.865013 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 13:46:50 crc kubenswrapper[4702]: E1125 13:46:50.865097 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 13:46:50 crc kubenswrapper[4702]: E1125 13:46:50.865156 4702 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 13:46:50 crc kubenswrapper[4702]: E1125 13:46:50.865328 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 13:46:51.3653033 +0000 UTC m=+20.527953750 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.865673 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.866850 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.866887 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.866941 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.867308 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.867363 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.867388 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.867328 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.867675 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.868014 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.868219 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.868038 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.868630 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.868687 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.868742 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.868902 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.868941 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.869205 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.869211 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.869385 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.869457 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.869602 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.869871 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.870117 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.870140 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.870163 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.870620 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.870873 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.870956 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.870963 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.871051 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.871222 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.871329 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.871504 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.871519 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.871590 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.871863 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.872096 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.872931 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.873323 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.875556 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.875582 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.875845 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.876055 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.876099 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.876127 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.876175 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.877134 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.878108 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.878119 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.878349 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.882873 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.890205 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.892465 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.893740 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.906395 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.908007 4702 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4" exitCode=255 Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.908050 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4"} Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.918222 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.919894 4702 scope.go:117] "RemoveContainer" containerID="4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.924813 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.930305 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.940735 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951072 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951446 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951495 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951535 4702 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951550 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951564 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951576 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951587 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951600 4702 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951612 4702 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951624 4702 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951635 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951633 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951646 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951756 4702 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951776 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951777 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951790 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951802 4702 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951813 4702 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951825 4702 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951837 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951849 4702 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951861 4702 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951873 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951884 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951895 4702 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951908 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951920 4702 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951933 4702 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951945 4702 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951957 4702 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951968 4702 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951980 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.951991 4702 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.952003 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.952016 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.952029 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.952041 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.952053 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.952065 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.952076 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.952088 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.952101 4702 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.952113 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.952124 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.952137 4702 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.952147 4702 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.952158 4702 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.952170 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.952182 4702 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.952194 4702 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.952206 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.952219 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.952247 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.952258 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.952269 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.952277 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.952286 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.952298 4702 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.952308 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.960852 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 13:46:50 crc kubenswrapper[4702]: I1125 13:46:50.978256 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.052641 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.060267 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 13:46:51 crc kubenswrapper[4702]: W1125 13:46:51.066615 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-cc3d64496a938b145b6fd82a7ffb9595987b29f747542e0ca8fdace052673f97 WatchSource:0}: Error finding container cc3d64496a938b145b6fd82a7ffb9595987b29f747542e0ca8fdace052673f97: Status 404 returned error can't find the container with id cc3d64496a938b145b6fd82a7ffb9595987b29f747542e0ca8fdace052673f97 Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.070452 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 13:46:51 crc kubenswrapper[4702]: W1125 13:46:51.083606 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-7d025e3d31c345ba0666eeda3a0734ee2b583f613f45799c9ae01f0ddddba30a WatchSource:0}: Error finding container 7d025e3d31c345ba0666eeda3a0734ee2b583f613f45799c9ae01f0ddddba30a: Status 404 returned error can't find the container with id 7d025e3d31c345ba0666eeda3a0734ee2b583f613f45799c9ae01f0ddddba30a Nov 25 13:46:51 crc kubenswrapper[4702]: W1125 13:46:51.086354 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-3dad67fd6d28f4aa128c872f45a4402d351419cfe6c8f7142e4370bf2b781667 WatchSource:0}: Error finding container 3dad67fd6d28f4aa128c872f45a4402d351419cfe6c8f7142e4370bf2b781667: Status 404 returned error can't find the container with id 3dad67fd6d28f4aa128c872f45a4402d351419cfe6c8f7142e4370bf2b781667 Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.354296 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.354388 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.354437 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:46:51 crc kubenswrapper[4702]: E1125 13:46:51.354482 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:46:52.354454277 +0000 UTC m=+21.517104727 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:46:51 crc kubenswrapper[4702]: E1125 13:46:51.354554 4702 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 13:46:51 crc kubenswrapper[4702]: E1125 13:46:51.354621 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 13:46:52.35460742 +0000 UTC m=+21.517257870 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 13:46:51 crc kubenswrapper[4702]: E1125 13:46:51.354616 4702 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 13:46:51 crc kubenswrapper[4702]: E1125 13:46:51.354748 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 13:46:52.354722033 +0000 UTC m=+21.517372483 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.455524 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.455595 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:46:51 crc kubenswrapper[4702]: E1125 13:46:51.455731 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 13:46:51 crc kubenswrapper[4702]: E1125 13:46:51.455762 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 13:46:51 crc kubenswrapper[4702]: E1125 13:46:51.455777 4702 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 13:46:51 crc kubenswrapper[4702]: E1125 13:46:51.455731 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 13:46:51 crc kubenswrapper[4702]: E1125 13:46:51.455856 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 13:46:51 crc kubenswrapper[4702]: E1125 13:46:51.455835 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 13:46:52.45581492 +0000 UTC m=+21.618465370 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 13:46:51 crc kubenswrapper[4702]: E1125 13:46:51.455867 4702 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 13:46:51 crc kubenswrapper[4702]: E1125 13:46:51.455903 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 13:46:52.455893332 +0000 UTC m=+21.618543782 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.804374 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.805077 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.806127 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.806787 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.807822 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.808369 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.809086 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.810171 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.810965 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.812109 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.812673 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.814120 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.814714 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.815324 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.816728 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.817315 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.817272 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:51Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.818403 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.818849 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.819633 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.820671 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.821158 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.822148 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.822631 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.823720 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.824130 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.824804 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.825916 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.826515 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.827559 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.828195 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.829296 4702 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.829423 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.831092 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.831944 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:51Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.832004 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.832570 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.833995 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.834759 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.835622 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.836299 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.837331 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.837772 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.838745 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.839389 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.840341 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.840800 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.841668 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.842168 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.843410 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.843959 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.845032 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.845579 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.846615 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.847181 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.847425 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:51Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.847660 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.860661 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:51Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.875670 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:51Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.890099 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:51Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.901396 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:51Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.911648 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"3dad67fd6d28f4aa128c872f45a4402d351419cfe6c8f7142e4370bf2b781667"} Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.913103 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a"} Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.913128 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd"} Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.913139 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"7d025e3d31c345ba0666eeda3a0734ee2b583f613f45799c9ae01f0ddddba30a"} Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.914349 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f"} Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.914371 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"cc3d64496a938b145b6fd82a7ffb9595987b29f747542e0ca8fdace052673f97"} Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.916037 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.923092 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef"} Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.928115 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:51Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.943460 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:51Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.958686 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:51Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:51 crc kubenswrapper[4702]: I1125 13:46:51.972685 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:51Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:52 crc kubenswrapper[4702]: I1125 13:46:52.000476 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:51Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:52 crc kubenswrapper[4702]: I1125 13:46:52.018777 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:52Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:52 crc kubenswrapper[4702]: I1125 13:46:52.034517 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:52Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:52 crc kubenswrapper[4702]: I1125 13:46:52.048421 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:52Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:52 crc kubenswrapper[4702]: I1125 13:46:52.063620 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:52Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:52 crc kubenswrapper[4702]: I1125 13:46:52.076665 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:52Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:52 crc kubenswrapper[4702]: I1125 13:46:52.091611 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:52Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:52 crc kubenswrapper[4702]: I1125 13:46:52.105516 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:52Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:52 crc kubenswrapper[4702]: I1125 13:46:52.118642 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:52Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:52 crc kubenswrapper[4702]: I1125 13:46:52.131644 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:52Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:52 crc kubenswrapper[4702]: I1125 13:46:52.366823 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:46:52 crc kubenswrapper[4702]: I1125 13:46:52.366925 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:46:52 crc kubenswrapper[4702]: E1125 13:46:52.367012 4702 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 13:46:52 crc kubenswrapper[4702]: E1125 13:46:52.367041 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:46:54.366990266 +0000 UTC m=+23.529640716 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:46:52 crc kubenswrapper[4702]: E1125 13:46:52.367087 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 13:46:54.367073748 +0000 UTC m=+23.529724198 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 13:46:52 crc kubenswrapper[4702]: I1125 13:46:52.367157 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:46:52 crc kubenswrapper[4702]: E1125 13:46:52.367339 4702 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 13:46:52 crc kubenswrapper[4702]: E1125 13:46:52.367424 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 13:46:54.367402076 +0000 UTC m=+23.530052606 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 13:46:52 crc kubenswrapper[4702]: I1125 13:46:52.467971 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:46:52 crc kubenswrapper[4702]: I1125 13:46:52.468040 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:46:52 crc kubenswrapper[4702]: E1125 13:46:52.468185 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 13:46:52 crc kubenswrapper[4702]: E1125 13:46:52.468203 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 13:46:52 crc kubenswrapper[4702]: E1125 13:46:52.468216 4702 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 13:46:52 crc kubenswrapper[4702]: E1125 13:46:52.468316 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 13:46:54.468293629 +0000 UTC m=+23.630944079 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 13:46:52 crc kubenswrapper[4702]: E1125 13:46:52.468329 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 13:46:52 crc kubenswrapper[4702]: E1125 13:46:52.468377 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 13:46:52 crc kubenswrapper[4702]: E1125 13:46:52.468393 4702 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 13:46:52 crc kubenswrapper[4702]: E1125 13:46:52.468479 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 13:46:54.468452903 +0000 UTC m=+23.631103433 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 13:46:52 crc kubenswrapper[4702]: I1125 13:46:52.800751 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:46:52 crc kubenswrapper[4702]: I1125 13:46:52.800813 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:46:52 crc kubenswrapper[4702]: I1125 13:46:52.800828 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:46:52 crc kubenswrapper[4702]: E1125 13:46:52.800915 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:46:52 crc kubenswrapper[4702]: E1125 13:46:52.800975 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:46:52 crc kubenswrapper[4702]: E1125 13:46:52.801112 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:46:52 crc kubenswrapper[4702]: I1125 13:46:52.924243 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:46:53 crc kubenswrapper[4702]: I1125 13:46:53.929282 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6"} Nov 25 13:46:53 crc kubenswrapper[4702]: I1125 13:46:53.949653 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:53Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:53 crc kubenswrapper[4702]: I1125 13:46:53.964682 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:53Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:53 crc kubenswrapper[4702]: I1125 13:46:53.980617 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:53Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:53 crc kubenswrapper[4702]: I1125 13:46:53.998124 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:53Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:54 crc kubenswrapper[4702]: I1125 13:46:54.016665 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:54Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:54 crc kubenswrapper[4702]: I1125 13:46:54.033277 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:54Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:54 crc kubenswrapper[4702]: I1125 13:46:54.051378 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:54Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:54 crc kubenswrapper[4702]: I1125 13:46:54.384340 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:46:54 crc kubenswrapper[4702]: I1125 13:46:54.384443 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:46:54 crc kubenswrapper[4702]: I1125 13:46:54.384472 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:46:54 crc kubenswrapper[4702]: E1125 13:46:54.384573 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:46:58.384538948 +0000 UTC m=+27.547189398 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:46:54 crc kubenswrapper[4702]: E1125 13:46:54.384581 4702 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 13:46:54 crc kubenswrapper[4702]: E1125 13:46:54.384661 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 13:46:58.384653481 +0000 UTC m=+27.547303921 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 13:46:54 crc kubenswrapper[4702]: E1125 13:46:54.384656 4702 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 13:46:54 crc kubenswrapper[4702]: E1125 13:46:54.384757 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 13:46:58.384733493 +0000 UTC m=+27.547383943 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 13:46:54 crc kubenswrapper[4702]: I1125 13:46:54.485101 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:46:54 crc kubenswrapper[4702]: I1125 13:46:54.485155 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:46:54 crc kubenswrapper[4702]: E1125 13:46:54.485296 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 13:46:54 crc kubenswrapper[4702]: E1125 13:46:54.485317 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 13:46:54 crc kubenswrapper[4702]: E1125 13:46:54.485328 4702 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 13:46:54 crc kubenswrapper[4702]: E1125 13:46:54.485388 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 13:46:58.485370688 +0000 UTC m=+27.648021138 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 13:46:54 crc kubenswrapper[4702]: E1125 13:46:54.485401 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 13:46:54 crc kubenswrapper[4702]: E1125 13:46:54.485464 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 13:46:54 crc kubenswrapper[4702]: E1125 13:46:54.485480 4702 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 13:46:54 crc kubenswrapper[4702]: E1125 13:46:54.485568 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 13:46:58.485536352 +0000 UTC m=+27.648186892 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 13:46:54 crc kubenswrapper[4702]: I1125 13:46:54.801016 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:46:54 crc kubenswrapper[4702]: I1125 13:46:54.801108 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:46:54 crc kubenswrapper[4702]: I1125 13:46:54.801107 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:46:54 crc kubenswrapper[4702]: E1125 13:46:54.801372 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:46:54 crc kubenswrapper[4702]: E1125 13:46:54.801418 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:46:54 crc kubenswrapper[4702]: E1125 13:46:54.801550 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:46:54 crc kubenswrapper[4702]: I1125 13:46:54.930752 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 25 13:46:54 crc kubenswrapper[4702]: I1125 13:46:54.944644 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:54Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:54 crc kubenswrapper[4702]: I1125 13:46:54.949080 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 25 13:46:54 crc kubenswrapper[4702]: I1125 13:46:54.955997 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 25 13:46:54 crc kubenswrapper[4702]: I1125 13:46:54.959553 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:54Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:54 crc kubenswrapper[4702]: I1125 13:46:54.976180 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:54Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:54 crc kubenswrapper[4702]: I1125 13:46:54.994497 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:54Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:55 crc kubenswrapper[4702]: I1125 13:46:55.016388 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:55 crc kubenswrapper[4702]: I1125 13:46:55.033403 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:55 crc kubenswrapper[4702]: I1125 13:46:55.051565 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:55 crc kubenswrapper[4702]: I1125 13:46:55.069636 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:55 crc kubenswrapper[4702]: I1125 13:46:55.086000 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:55 crc kubenswrapper[4702]: I1125 13:46:55.108512 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:55 crc kubenswrapper[4702]: I1125 13:46:55.122339 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:55 crc kubenswrapper[4702]: I1125 13:46:55.137398 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:55 crc kubenswrapper[4702]: I1125 13:46:55.151769 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:55 crc kubenswrapper[4702]: I1125 13:46:55.165400 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:55 crc kubenswrapper[4702]: I1125 13:46:55.189861 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:55 crc kubenswrapper[4702]: I1125 13:46:55.834583 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-5mp5w"] Nov 25 13:46:55 crc kubenswrapper[4702]: I1125 13:46:55.834990 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-5mp5w" Nov 25 13:46:55 crc kubenswrapper[4702]: W1125 13:46:55.842045 4702 reflector.go:561] object-"openshift-dns"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-dns": no relationship found between node 'crc' and this object Nov 25 13:46:55 crc kubenswrapper[4702]: E1125 13:46:55.842109 4702 reflector.go:158] "Unhandled Error" err="object-\"openshift-dns\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-dns\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 13:46:55 crc kubenswrapper[4702]: W1125 13:46:55.847715 4702 reflector.go:561] object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": failed to list *v1.Secret: secrets "node-resolver-dockercfg-kz9s7" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-dns": no relationship found between node 'crc' and this object Nov 25 13:46:55 crc kubenswrapper[4702]: E1125 13:46:55.847783 4702 reflector.go:158] "Unhandled Error" err="object-\"openshift-dns\"/\"node-resolver-dockercfg-kz9s7\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"node-resolver-dockercfg-kz9s7\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-dns\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 13:46:55 crc kubenswrapper[4702]: I1125 13:46:55.847732 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 25 13:46:55 crc kubenswrapper[4702]: I1125 13:46:55.897134 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnsck\" (UniqueName: \"kubernetes.io/projected/83c1c667-d798-4bdc-83d0-25c77a44fc43-kube-api-access-vnsck\") pod \"node-resolver-5mp5w\" (UID: \"83c1c667-d798-4bdc-83d0-25c77a44fc43\") " pod="openshift-dns/node-resolver-5mp5w" Nov 25 13:46:55 crc kubenswrapper[4702]: I1125 13:46:55.905688 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/83c1c667-d798-4bdc-83d0-25c77a44fc43-hosts-file\") pod \"node-resolver-5mp5w\" (UID: \"83c1c667-d798-4bdc-83d0-25c77a44fc43\") " pod="openshift-dns/node-resolver-5mp5w" Nov 25 13:46:55 crc kubenswrapper[4702]: I1125 13:46:55.906149 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:55 crc kubenswrapper[4702]: I1125 13:46:55.935876 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:55 crc kubenswrapper[4702]: I1125 13:46:55.961175 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:55 crc kubenswrapper[4702]: I1125 13:46:55.975090 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:55 crc kubenswrapper[4702]: I1125 13:46:55.988201 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.003361 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.006213 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/83c1c667-d798-4bdc-83d0-25c77a44fc43-hosts-file\") pod \"node-resolver-5mp5w\" (UID: \"83c1c667-d798-4bdc-83d0-25c77a44fc43\") " pod="openshift-dns/node-resolver-5mp5w" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.006336 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnsck\" (UniqueName: \"kubernetes.io/projected/83c1c667-d798-4bdc-83d0-25c77a44fc43-kube-api-access-vnsck\") pod \"node-resolver-5mp5w\" (UID: \"83c1c667-d798-4bdc-83d0-25c77a44fc43\") " pod="openshift-dns/node-resolver-5mp5w" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.006342 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/83c1c667-d798-4bdc-83d0-25c77a44fc43-hosts-file\") pod \"node-resolver-5mp5w\" (UID: \"83c1c667-d798-4bdc-83d0-25c77a44fc43\") " pod="openshift-dns/node-resolver-5mp5w" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.013643 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.028248 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.043891 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.254137 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-qqr76"] Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.254606 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.256682 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.256682 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.257038 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.257367 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.259379 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.273267 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.290976 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.306782 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.309261 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3c491818-b0bb-4d82-a031-96e2dfac8c27-mcd-auth-proxy-config\") pod \"machine-config-daemon-qqr76\" (UID: \"3c491818-b0bb-4d82-a031-96e2dfac8c27\") " pod="openshift-machine-config-operator/machine-config-daemon-qqr76" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.309331 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgszd\" (UniqueName: \"kubernetes.io/projected/3c491818-b0bb-4d82-a031-96e2dfac8c27-kube-api-access-vgszd\") pod \"machine-config-daemon-qqr76\" (UID: \"3c491818-b0bb-4d82-a031-96e2dfac8c27\") " pod="openshift-machine-config-operator/machine-config-daemon-qqr76" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.309508 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3c491818-b0bb-4d82-a031-96e2dfac8c27-proxy-tls\") pod \"machine-config-daemon-qqr76\" (UID: \"3c491818-b0bb-4d82-a031-96e2dfac8c27\") " pod="openshift-machine-config-operator/machine-config-daemon-qqr76" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.309660 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/3c491818-b0bb-4d82-a031-96e2dfac8c27-rootfs\") pod \"machine-config-daemon-qqr76\" (UID: \"3c491818-b0bb-4d82-a031-96e2dfac8c27\") " pod="openshift-machine-config-operator/machine-config-daemon-qqr76" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.327809 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.342790 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.357180 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.379813 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.391447 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.403705 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.410417 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3c491818-b0bb-4d82-a031-96e2dfac8c27-mcd-auth-proxy-config\") pod \"machine-config-daemon-qqr76\" (UID: \"3c491818-b0bb-4d82-a031-96e2dfac8c27\") " pod="openshift-machine-config-operator/machine-config-daemon-qqr76" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.410529 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgszd\" (UniqueName: \"kubernetes.io/projected/3c491818-b0bb-4d82-a031-96e2dfac8c27-kube-api-access-vgszd\") pod \"machine-config-daemon-qqr76\" (UID: \"3c491818-b0bb-4d82-a031-96e2dfac8c27\") " pod="openshift-machine-config-operator/machine-config-daemon-qqr76" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.411022 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3c491818-b0bb-4d82-a031-96e2dfac8c27-proxy-tls\") pod \"machine-config-daemon-qqr76\" (UID: \"3c491818-b0bb-4d82-a031-96e2dfac8c27\") " pod="openshift-machine-config-operator/machine-config-daemon-qqr76" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.411456 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3c491818-b0bb-4d82-a031-96e2dfac8c27-mcd-auth-proxy-config\") pod \"machine-config-daemon-qqr76\" (UID: \"3c491818-b0bb-4d82-a031-96e2dfac8c27\") " pod="openshift-machine-config-operator/machine-config-daemon-qqr76" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.411643 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/3c491818-b0bb-4d82-a031-96e2dfac8c27-rootfs\") pod \"machine-config-daemon-qqr76\" (UID: \"3c491818-b0bb-4d82-a031-96e2dfac8c27\") " pod="openshift-machine-config-operator/machine-config-daemon-qqr76" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.411701 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/3c491818-b0bb-4d82-a031-96e2dfac8c27-rootfs\") pod \"machine-config-daemon-qqr76\" (UID: \"3c491818-b0bb-4d82-a031-96e2dfac8c27\") " pod="openshift-machine-config-operator/machine-config-daemon-qqr76" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.418099 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3c491818-b0bb-4d82-a031-96e2dfac8c27-proxy-tls\") pod \"machine-config-daemon-qqr76\" (UID: \"3c491818-b0bb-4d82-a031-96e2dfac8c27\") " pod="openshift-machine-config-operator/machine-config-daemon-qqr76" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.422103 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.436358 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgszd\" (UniqueName: \"kubernetes.io/projected/3c491818-b0bb-4d82-a031-96e2dfac8c27-kube-api-access-vgszd\") pod \"machine-config-daemon-qqr76\" (UID: \"3c491818-b0bb-4d82-a031-96e2dfac8c27\") " pod="openshift-machine-config-operator/machine-config-daemon-qqr76" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.567181 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.580532 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 13:46:56 crc kubenswrapper[4702]: W1125 13:46:56.593787 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c491818_b0bb_4d82_a031_96e2dfac8c27.slice/crio-03941b99e86fd909aa40a5dea80f90772a5a526a953163447da92310a510d315 WatchSource:0}: Error finding container 03941b99e86fd909aa40a5dea80f90772a5a526a953163447da92310a510d315: Status 404 returned error can't find the container with id 03941b99e86fd909aa40a5dea80f90772a5a526a953163447da92310a510d315 Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.599487 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.600553 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.617772 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.635678 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.636738 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.650767 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.652659 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-gd2xs"] Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.653051 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.653949 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-wcs7x"] Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.654686 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.659615 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-p9c5v"] Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.660658 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: W1125 13:46:56.662123 4702 reflector.go:561] object-"openshift-multus"/"cni-copy-resources": failed to list *v1.ConfigMap: configmaps "cni-copy-resources" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Nov 25 13:46:56 crc kubenswrapper[4702]: W1125 13:46:56.662143 4702 reflector.go:561] object-"openshift-multus"/"multus-daemon-config": failed to list *v1.ConfigMap: configmaps "multus-daemon-config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Nov 25 13:46:56 crc kubenswrapper[4702]: E1125 13:46:56.662170 4702 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"cni-copy-resources\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"cni-copy-resources\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 13:46:56 crc kubenswrapper[4702]: W1125 13:46:56.663178 4702 reflector.go:561] object-"openshift-multus"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Nov 25 13:46:56 crc kubenswrapper[4702]: W1125 13:46:56.662123 4702 reflector.go:561] object-"openshift-multus"/"default-dockercfg-2q5b6": failed to list *v1.Secret: secrets "default-dockercfg-2q5b6" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Nov 25 13:46:56 crc kubenswrapper[4702]: E1125 13:46:56.663301 4702 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"default-dockercfg-2q5b6\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"default-dockercfg-2q5b6\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 13:46:56 crc kubenswrapper[4702]: W1125 13:46:56.663314 4702 reflector.go:561] object-"openshift-multus"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Nov 25 13:46:56 crc kubenswrapper[4702]: E1125 13:46:56.663350 4702 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 13:46:56 crc kubenswrapper[4702]: W1125 13:46:56.663352 4702 reflector.go:561] object-"openshift-multus"/"default-cni-sysctl-allowlist": failed to list *v1.ConfigMap: configmaps "default-cni-sysctl-allowlist" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Nov 25 13:46:56 crc kubenswrapper[4702]: E1125 13:46:56.663313 4702 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 13:46:56 crc kubenswrapper[4702]: E1125 13:46:56.663404 4702 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"default-cni-sysctl-allowlist\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"default-cni-sysctl-allowlist\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 13:46:56 crc kubenswrapper[4702]: W1125 13:46:56.663427 4702 reflector.go:561] object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": failed to list *v1.Secret: secrets "multus-ancillary-tools-dockercfg-vnmsz" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Nov 25 13:46:56 crc kubenswrapper[4702]: E1125 13:46:56.663528 4702 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"multus-ancillary-tools-dockercfg-vnmsz\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"multus-ancillary-tools-dockercfg-vnmsz\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 13:46:56 crc kubenswrapper[4702]: E1125 13:46:56.662194 4702 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"multus-daemon-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"multus-daemon-config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.670718 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.671095 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.671272 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.671463 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.671577 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.671693 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.671806 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.681858 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.715042 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-node-log\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.715381 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-host-var-lib-kubelet\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.715515 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b142b34b-79ab-4995-9578-d8fdf04e1fba-system-cni-dir\") pod \"multus-additional-cni-plugins-wcs7x\" (UID: \"b142b34b-79ab-4995-9578-d8fdf04e1fba\") " pod="openshift-multus/multus-additional-cni-plugins-wcs7x" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.715664 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-run-netns\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.715762 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-cnibin\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.715782 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-os-release\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.715880 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-host-var-lib-cni-multus\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.715949 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.715983 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdztt\" (UniqueName: \"kubernetes.io/projected/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-kube-api-access-hdztt\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.716027 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-cni-bin\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.716057 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-system-cni-dir\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.716079 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-host-var-lib-cni-bin\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.716106 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sj98\" (UniqueName: \"kubernetes.io/projected/9e735a90-cf24-47de-b0e6-57d27a971b21-kube-api-access-6sj98\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.716133 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-ovn-node-metrics-cert\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.716163 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-host-run-netns\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.716188 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-run-ovn\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.716210 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-env-overrides\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.716258 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-multus-socket-dir-parent\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.716335 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/9e735a90-cf24-47de-b0e6-57d27a971b21-multus-daemon-config\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.716381 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-slash\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.716432 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-etc-openvswitch\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.716451 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-multus-conf-dir\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.716514 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-hostroot\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.716587 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b142b34b-79ab-4995-9578-d8fdf04e1fba-os-release\") pod \"multus-additional-cni-plugins-wcs7x\" (UID: \"b142b34b-79ab-4995-9578-d8fdf04e1fba\") " pod="openshift-multus/multus-additional-cni-plugins-wcs7x" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.716629 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-host-run-multus-certs\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.716647 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-etc-kubernetes\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.716673 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-cni-netd\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.716690 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmrxm\" (UniqueName: \"kubernetes.io/projected/b142b34b-79ab-4995-9578-d8fdf04e1fba-kube-api-access-fmrxm\") pod \"multus-additional-cni-plugins-wcs7x\" (UID: \"b142b34b-79ab-4995-9578-d8fdf04e1fba\") " pod="openshift-multus/multus-additional-cni-plugins-wcs7x" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.716710 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-systemd-units\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.716731 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-run-systemd\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.716748 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-run-openvswitch\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.716763 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-multus-cni-dir\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.716780 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9e735a90-cf24-47de-b0e6-57d27a971b21-cni-binary-copy\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.716802 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-host-run-k8s-cni-cncf-io\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.716882 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b142b34b-79ab-4995-9578-d8fdf04e1fba-cnibin\") pod \"multus-additional-cni-plugins-wcs7x\" (UID: \"b142b34b-79ab-4995-9578-d8fdf04e1fba\") " pod="openshift-multus/multus-additional-cni-plugins-wcs7x" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.716964 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b142b34b-79ab-4995-9578-d8fdf04e1fba-cni-binary-copy\") pod \"multus-additional-cni-plugins-wcs7x\" (UID: \"b142b34b-79ab-4995-9578-d8fdf04e1fba\") " pod="openshift-multus/multus-additional-cni-plugins-wcs7x" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.717021 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-kubelet\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.717045 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-var-lib-openvswitch\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.717074 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-ovnkube-config\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.717100 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-ovnkube-script-lib\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.717124 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/b142b34b-79ab-4995-9578-d8fdf04e1fba-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-wcs7x\" (UID: \"b142b34b-79ab-4995-9578-d8fdf04e1fba\") " pod="openshift-multus/multus-additional-cni-plugins-wcs7x" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.717151 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-log-socket\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.717176 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-run-ovn-kubernetes\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.717279 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b142b34b-79ab-4995-9578-d8fdf04e1fba-tuning-conf-dir\") pod \"multus-additional-cni-plugins-wcs7x\" (UID: \"b142b34b-79ab-4995-9578-d8fdf04e1fba\") " pod="openshift-multus/multus-additional-cni-plugins-wcs7x" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.723027 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.742113 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.755555 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.778397 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.794468 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.800511 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.800553 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.800522 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:46:56 crc kubenswrapper[4702]: E1125 13:46:56.800682 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:46:56 crc kubenswrapper[4702]: E1125 13:46:56.800740 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:46:56 crc kubenswrapper[4702]: E1125 13:46:56.800870 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.812059 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.817934 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b142b34b-79ab-4995-9578-d8fdf04e1fba-os-release\") pod \"multus-additional-cni-plugins-wcs7x\" (UID: \"b142b34b-79ab-4995-9578-d8fdf04e1fba\") " pod="openshift-multus/multus-additional-cni-plugins-wcs7x" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.817986 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-etc-kubernetes\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818015 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-host-run-multus-certs\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818040 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmrxm\" (UniqueName: \"kubernetes.io/projected/b142b34b-79ab-4995-9578-d8fdf04e1fba-kube-api-access-fmrxm\") pod \"multus-additional-cni-plugins-wcs7x\" (UID: \"b142b34b-79ab-4995-9578-d8fdf04e1fba\") " pod="openshift-multus/multus-additional-cni-plugins-wcs7x" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818066 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-systemd-units\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818082 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-etc-kubernetes\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818092 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-run-systemd\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818147 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-run-openvswitch\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818151 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-host-run-multus-certs\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818196 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-cni-netd\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818119 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-run-systemd\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818171 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-cni-netd\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818260 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-systemd-units\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818289 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-multus-cni-dir\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818322 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9e735a90-cf24-47de-b0e6-57d27a971b21-cni-binary-copy\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818347 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-host-run-k8s-cni-cncf-io\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818295 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-run-openvswitch\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818374 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-var-lib-openvswitch\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818403 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b142b34b-79ab-4995-9578-d8fdf04e1fba-cnibin\") pod \"multus-additional-cni-plugins-wcs7x\" (UID: \"b142b34b-79ab-4995-9578-d8fdf04e1fba\") " pod="openshift-multus/multus-additional-cni-plugins-wcs7x" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818411 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-host-run-k8s-cni-cncf-io\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818364 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b142b34b-79ab-4995-9578-d8fdf04e1fba-os-release\") pod \"multus-additional-cni-plugins-wcs7x\" (UID: \"b142b34b-79ab-4995-9578-d8fdf04e1fba\") " pod="openshift-multus/multus-additional-cni-plugins-wcs7x" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818424 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b142b34b-79ab-4995-9578-d8fdf04e1fba-cni-binary-copy\") pod \"multus-additional-cni-plugins-wcs7x\" (UID: \"b142b34b-79ab-4995-9578-d8fdf04e1fba\") " pod="openshift-multus/multus-additional-cni-plugins-wcs7x" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818489 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b142b34b-79ab-4995-9578-d8fdf04e1fba-cnibin\") pod \"multus-additional-cni-plugins-wcs7x\" (UID: \"b142b34b-79ab-4995-9578-d8fdf04e1fba\") " pod="openshift-multus/multus-additional-cni-plugins-wcs7x" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818479 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-var-lib-openvswitch\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818522 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-kubelet\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818547 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-ovnkube-config\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818561 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-multus-cni-dir\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818574 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-ovnkube-script-lib\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818551 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-kubelet\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818592 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/b142b34b-79ab-4995-9578-d8fdf04e1fba-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-wcs7x\" (UID: \"b142b34b-79ab-4995-9578-d8fdf04e1fba\") " pod="openshift-multus/multus-additional-cni-plugins-wcs7x" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818609 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-log-socket\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818627 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-run-ovn-kubernetes\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818649 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b142b34b-79ab-4995-9578-d8fdf04e1fba-tuning-conf-dir\") pod \"multus-additional-cni-plugins-wcs7x\" (UID: \"b142b34b-79ab-4995-9578-d8fdf04e1fba\") " pod="openshift-multus/multus-additional-cni-plugins-wcs7x" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818668 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-node-log\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818685 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-host-var-lib-kubelet\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818706 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b142b34b-79ab-4995-9578-d8fdf04e1fba-system-cni-dir\") pod \"multus-additional-cni-plugins-wcs7x\" (UID: \"b142b34b-79ab-4995-9578-d8fdf04e1fba\") " pod="openshift-multus/multus-additional-cni-plugins-wcs7x" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818728 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-run-netns\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818760 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-cnibin\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818790 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-os-release\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818812 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-host-var-lib-cni-multus\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818831 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-cni-bin\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818848 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818865 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdztt\" (UniqueName: \"kubernetes.io/projected/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-kube-api-access-hdztt\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818880 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-system-cni-dir\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818895 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-host-var-lib-cni-bin\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818913 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-host-run-netns\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818934 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sj98\" (UniqueName: \"kubernetes.io/projected/9e735a90-cf24-47de-b0e6-57d27a971b21-kube-api-access-6sj98\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818954 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-ovn-node-metrics-cert\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.818974 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-multus-socket-dir-parent\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.819003 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/9e735a90-cf24-47de-b0e6-57d27a971b21-multus-daemon-config\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.819019 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-slash\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.819033 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-run-ovn\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.819048 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-env-overrides\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.819081 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-etc-openvswitch\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.819107 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-hostroot\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.819123 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-multus-conf-dir\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.819191 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-multus-conf-dir\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.819283 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-log-socket\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.819304 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-run-ovn-kubernetes\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.819398 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-ovnkube-script-lib\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.819398 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-ovnkube-config\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.819490 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-host-var-lib-cni-bin\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.819521 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-system-cni-dir\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.819527 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-host-run-netns\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.819568 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-slash\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.819604 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-host-var-lib-kubelet\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.819607 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b142b34b-79ab-4995-9578-d8fdf04e1fba-system-cni-dir\") pod \"multus-additional-cni-plugins-wcs7x\" (UID: \"b142b34b-79ab-4995-9578-d8fdf04e1fba\") " pod="openshift-multus/multus-additional-cni-plugins-wcs7x" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.819644 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-run-ovn\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.819680 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-run-netns\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.819725 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-cnibin\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.819777 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-os-release\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.819808 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-host-var-lib-cni-multus\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.819841 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-cni-bin\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.819873 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.820066 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-multus-socket-dir-parent\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.819573 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-node-log\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.820106 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-env-overrides\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.820103 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/9e735a90-cf24-47de-b0e6-57d27a971b21-hostroot\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.820176 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-etc-openvswitch\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.824078 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-ovn-node-metrics-cert\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.828881 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.837537 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdztt\" (UniqueName: \"kubernetes.io/projected/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-kube-api-access-hdztt\") pod \"ovnkube-node-p9c5v\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.843517 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.863202 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.875101 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.890382 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.890618 4702 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.892306 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.892346 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.892357 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.892481 4702 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.899073 4702 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.899425 4702 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.900771 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.900832 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.900845 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.900864 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.900878 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:56Z","lastTransitionTime":"2025-11-25T13:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.910941 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: E1125 13:46:56.918093 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"96cc93a3-66f3-477f-bbc8-ef3a984ed406\\\",\\\"systemUUID\\\":\\\"bb366f18-cc3c-451f-9d6a-c0cc41864b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.922831 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.922882 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.922898 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.922920 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.922933 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:56Z","lastTransitionTime":"2025-11-25T13:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.929549 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.937865 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" event={"ID":"3c491818-b0bb-4d82-a031-96e2dfac8c27","Type":"ContainerStarted","Data":"03941b99e86fd909aa40a5dea80f90772a5a526a953163447da92310a510d315"} Nov 25 13:46:56 crc kubenswrapper[4702]: E1125 13:46:56.938282 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"96cc93a3-66f3-477f-bbc8-ef3a984ed406\\\",\\\"systemUUID\\\":\\\"bb366f18-cc3c-451f-9d6a-c0cc41864b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.941583 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.941718 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.941792 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.941866 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.941926 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:56Z","lastTransitionTime":"2025-11-25T13:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.948082 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: E1125 13:46:56.948616 4702 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 13:46:56 crc kubenswrapper[4702]: E1125 13:46:56.957879 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"96cc93a3-66f3-477f-bbc8-ef3a984ed406\\\",\\\"systemUUID\\\":\\\"bb366f18-cc3c-451f-9d6a-c0cc41864b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.961708 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.961737 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.961746 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.961761 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.961771 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:56Z","lastTransitionTime":"2025-11-25T13:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.961708 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: E1125 13:46:56.972289 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"96cc93a3-66f3-477f-bbc8-ef3a984ed406\\\",\\\"systemUUID\\\":\\\"bb366f18-cc3c-451f-9d6a-c0cc41864b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.973871 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.976562 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.976718 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.976789 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.976857 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.976924 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:56Z","lastTransitionTime":"2025-11-25T13:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.983568 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: E1125 13:46:56.989764 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"96cc93a3-66f3-477f-bbc8-ef3a984ed406\\\",\\\"systemUUID\\\":\\\"bb366f18-cc3c-451f-9d6a-c0cc41864b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: E1125 13:46:56.989950 4702 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.991883 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.991944 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.991955 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.991973 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.992008 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:56Z","lastTransitionTime":"2025-11-25T13:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.995692 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.996573 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:56 crc kubenswrapper[4702]: I1125 13:46:56.999903 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b142b34b-79ab-4995-9578-d8fdf04e1fba-tuning-conf-dir\") pod \"multus-additional-cni-plugins-wcs7x\" (UID: \"b142b34b-79ab-4995-9578-d8fdf04e1fba\") " pod="openshift-multus/multus-additional-cni-plugins-wcs7x" Nov 25 13:46:57 crc kubenswrapper[4702]: W1125 13:46:57.010683 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c8a3234_01d3_48a9_a5b4_acb46b4218dc.slice/crio-7732801e549701b7f73321af4fb72c468f45cae12714478a77d6aeacada97801 WatchSource:0}: Error finding container 7732801e549701b7f73321af4fb72c468f45cae12714478a77d6aeacada97801: Status 404 returned error can't find the container with id 7732801e549701b7f73321af4fb72c468f45cae12714478a77d6aeacada97801 Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.018033 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:57Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:57 crc kubenswrapper[4702]: E1125 13:46:57.021822 4702 projected.go:288] Couldn't get configMap openshift-dns/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Nov 25 13:46:57 crc kubenswrapper[4702]: E1125 13:46:57.021869 4702 projected.go:194] Error preparing data for projected volume kube-api-access-vnsck for pod openshift-dns/node-resolver-5mp5w: failed to sync configmap cache: timed out waiting for the condition Nov 25 13:46:57 crc kubenswrapper[4702]: E1125 13:46:57.021928 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/83c1c667-d798-4bdc-83d0-25c77a44fc43-kube-api-access-vnsck podName:83c1c667-d798-4bdc-83d0-25c77a44fc43 nodeName:}" failed. No retries permitted until 2025-11-25 13:46:57.521910394 +0000 UTC m=+26.684560844 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-vnsck" (UniqueName: "kubernetes.io/projected/83c1c667-d798-4bdc-83d0-25c77a44fc43-kube-api-access-vnsck") pod "node-resolver-5mp5w" (UID: "83c1c667-d798-4bdc-83d0-25c77a44fc43") : failed to sync configmap cache: timed out waiting for the condition Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.073697 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.094100 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.094134 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.094143 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.094159 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.094170 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:57Z","lastTransitionTime":"2025-11-25T13:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.156767 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.196366 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.196404 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.196412 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.196426 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.196437 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:57Z","lastTransitionTime":"2025-11-25T13:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.299789 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.299828 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.299840 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.299857 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.299872 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:57Z","lastTransitionTime":"2025-11-25T13:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.401668 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.401712 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.401721 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.401735 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.401746 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:57Z","lastTransitionTime":"2025-11-25T13:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.504082 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.504119 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.504127 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.504143 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.504154 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:57Z","lastTransitionTime":"2025-11-25T13:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.527574 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnsck\" (UniqueName: \"kubernetes.io/projected/83c1c667-d798-4bdc-83d0-25c77a44fc43-kube-api-access-vnsck\") pod \"node-resolver-5mp5w\" (UID: \"83c1c667-d798-4bdc-83d0-25c77a44fc43\") " pod="openshift-dns/node-resolver-5mp5w" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.530216 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.531139 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnsck\" (UniqueName: \"kubernetes.io/projected/83c1c667-d798-4bdc-83d0-25c77a44fc43-kube-api-access-vnsck\") pod \"node-resolver-5mp5w\" (UID: \"83c1c667-d798-4bdc-83d0-25c77a44fc43\") " pod="openshift-dns/node-resolver-5mp5w" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.606054 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.606098 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.606109 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.606126 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.606140 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:57Z","lastTransitionTime":"2025-11-25T13:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.647540 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-5mp5w" Nov 25 13:46:57 crc kubenswrapper[4702]: W1125 13:46:57.660831 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod83c1c667_d798_4bdc_83d0_25c77a44fc43.slice/crio-9e0876d09d5f2813e718542fa66e21a7930fa320d7356632e79469caa161dc57 WatchSource:0}: Error finding container 9e0876d09d5f2813e718542fa66e21a7930fa320d7356632e79469caa161dc57: Status 404 returned error can't find the container with id 9e0876d09d5f2813e718542fa66e21a7930fa320d7356632e79469caa161dc57 Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.710810 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.710880 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.710943 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.710960 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.710981 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:57Z","lastTransitionTime":"2025-11-25T13:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.813357 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.813407 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.813419 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.813439 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.813452 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:57Z","lastTransitionTime":"2025-11-25T13:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:57 crc kubenswrapper[4702]: E1125 13:46:57.818756 4702 configmap.go:193] Couldn't get configMap openshift-multus/cni-copy-resources: failed to sync configmap cache: timed out waiting for the condition Nov 25 13:46:57 crc kubenswrapper[4702]: E1125 13:46:57.818824 4702 configmap.go:193] Couldn't get configMap openshift-multus/cni-copy-resources: failed to sync configmap cache: timed out waiting for the condition Nov 25 13:46:57 crc kubenswrapper[4702]: E1125 13:46:57.818859 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/9e735a90-cf24-47de-b0e6-57d27a971b21-cni-binary-copy podName:9e735a90-cf24-47de-b0e6-57d27a971b21 nodeName:}" failed. No retries permitted until 2025-11-25 13:46:58.318834424 +0000 UTC m=+27.481484874 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cni-binary-copy" (UniqueName: "kubernetes.io/configmap/9e735a90-cf24-47de-b0e6-57d27a971b21-cni-binary-copy") pod "multus-gd2xs" (UID: "9e735a90-cf24-47de-b0e6-57d27a971b21") : failed to sync configmap cache: timed out waiting for the condition Nov 25 13:46:57 crc kubenswrapper[4702]: E1125 13:46:57.818908 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b142b34b-79ab-4995-9578-d8fdf04e1fba-cni-binary-copy podName:b142b34b-79ab-4995-9578-d8fdf04e1fba nodeName:}" failed. No retries permitted until 2025-11-25 13:46:58.318883005 +0000 UTC m=+27.481533525 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cni-binary-copy" (UniqueName: "kubernetes.io/configmap/b142b34b-79ab-4995-9578-d8fdf04e1fba-cni-binary-copy") pod "multus-additional-cni-plugins-wcs7x" (UID: "b142b34b-79ab-4995-9578-d8fdf04e1fba") : failed to sync configmap cache: timed out waiting for the condition Nov 25 13:46:57 crc kubenswrapper[4702]: E1125 13:46:57.819874 4702 configmap.go:193] Couldn't get configMap openshift-multus/default-cni-sysctl-allowlist: failed to sync configmap cache: timed out waiting for the condition Nov 25 13:46:57 crc kubenswrapper[4702]: E1125 13:46:57.819909 4702 configmap.go:193] Couldn't get configMap openshift-multus/multus-daemon-config: failed to sync configmap cache: timed out waiting for the condition Nov 25 13:46:57 crc kubenswrapper[4702]: E1125 13:46:57.819932 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b142b34b-79ab-4995-9578-d8fdf04e1fba-cni-sysctl-allowlist podName:b142b34b-79ab-4995-9578-d8fdf04e1fba nodeName:}" failed. No retries permitted until 2025-11-25 13:46:58.319918791 +0000 UTC m=+27.482569331 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cni-sysctl-allowlist" (UniqueName: "kubernetes.io/configmap/b142b34b-79ab-4995-9578-d8fdf04e1fba-cni-sysctl-allowlist") pod "multus-additional-cni-plugins-wcs7x" (UID: "b142b34b-79ab-4995-9578-d8fdf04e1fba") : failed to sync configmap cache: timed out waiting for the condition Nov 25 13:46:57 crc kubenswrapper[4702]: E1125 13:46:57.819950 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/9e735a90-cf24-47de-b0e6-57d27a971b21-multus-daemon-config podName:9e735a90-cf24-47de-b0e6-57d27a971b21 nodeName:}" failed. No retries permitted until 2025-11-25 13:46:58.319941102 +0000 UTC m=+27.482591552 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "multus-daemon-config" (UniqueName: "kubernetes.io/configmap/9e735a90-cf24-47de-b0e6-57d27a971b21-multus-daemon-config") pod "multus-gd2xs" (UID: "9e735a90-cf24-47de-b0e6-57d27a971b21") : failed to sync configmap cache: timed out waiting for the condition Nov 25 13:46:57 crc kubenswrapper[4702]: E1125 13:46:57.834599 4702 projected.go:288] Couldn't get configMap openshift-multus/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Nov 25 13:46:57 crc kubenswrapper[4702]: E1125 13:46:57.836324 4702 projected.go:288] Couldn't get configMap openshift-multus/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.837822 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.852701 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.916049 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.916083 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.916091 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.916104 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.916114 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:57Z","lastTransitionTime":"2025-11-25T13:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.930442 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.941916 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-5mp5w" event={"ID":"83c1c667-d798-4bdc-83d0-25c77a44fc43","Type":"ContainerStarted","Data":"c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb"} Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.941983 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-5mp5w" event={"ID":"83c1c667-d798-4bdc-83d0-25c77a44fc43","Type":"ContainerStarted","Data":"9e0876d09d5f2813e718542fa66e21a7930fa320d7356632e79469caa161dc57"} Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.944616 4702 generic.go:334] "Generic (PLEG): container finished" podID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerID="d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8" exitCode=0 Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.944673 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" event={"ID":"3c8a3234-01d3-48a9-a5b4-acb46b4218dc","Type":"ContainerDied","Data":"d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8"} Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.944701 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" event={"ID":"3c8a3234-01d3-48a9-a5b4-acb46b4218dc","Type":"ContainerStarted","Data":"7732801e549701b7f73321af4fb72c468f45cae12714478a77d6aeacada97801"} Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.946478 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" event={"ID":"3c491818-b0bb-4d82-a031-96e2dfac8c27","Type":"ContainerStarted","Data":"a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a"} Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.946517 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" event={"ID":"3c491818-b0bb-4d82-a031-96e2dfac8c27","Type":"ContainerStarted","Data":"1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a"} Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.958185 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.959696 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:57Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:57 crc kubenswrapper[4702]: E1125 13:46:57.964751 4702 projected.go:194] Error preparing data for projected volume kube-api-access-6sj98 for pod openshift-multus/multus-gd2xs: failed to sync configmap cache: timed out waiting for the condition Nov 25 13:46:57 crc kubenswrapper[4702]: E1125 13:46:57.965014 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9e735a90-cf24-47de-b0e6-57d27a971b21-kube-api-access-6sj98 podName:9e735a90-cf24-47de-b0e6-57d27a971b21 nodeName:}" failed. No retries permitted until 2025-11-25 13:46:58.464978206 +0000 UTC m=+27.627628656 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-6sj98" (UniqueName: "kubernetes.io/projected/9e735a90-cf24-47de-b0e6-57d27a971b21-kube-api-access-6sj98") pod "multus-gd2xs" (UID: "9e735a90-cf24-47de-b0e6-57d27a971b21") : failed to sync configmap cache: timed out waiting for the condition Nov 25 13:46:57 crc kubenswrapper[4702]: E1125 13:46:57.967086 4702 projected.go:194] Error preparing data for projected volume kube-api-access-fmrxm for pod openshift-multus/multus-additional-cni-plugins-wcs7x: failed to sync configmap cache: timed out waiting for the condition Nov 25 13:46:57 crc kubenswrapper[4702]: E1125 13:46:57.967194 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b142b34b-79ab-4995-9578-d8fdf04e1fba-kube-api-access-fmrxm podName:b142b34b-79ab-4995-9578-d8fdf04e1fba nodeName:}" failed. No retries permitted until 2025-11-25 13:46:58.467166752 +0000 UTC m=+27.629817272 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-fmrxm" (UniqueName: "kubernetes.io/projected/b142b34b-79ab-4995-9578-d8fdf04e1fba-kube-api-access-fmrxm") pod "multus-additional-cni-plugins-wcs7x" (UID: "b142b34b-79ab-4995-9578-d8fdf04e1fba") : failed to sync configmap cache: timed out waiting for the condition Nov 25 13:46:57 crc kubenswrapper[4702]: I1125 13:46:57.974180 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:57Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.000513 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:57Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.015634 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.019007 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.019033 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.019041 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.019055 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.019064 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:58Z","lastTransitionTime":"2025-11-25T13:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.030489 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.041943 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.043201 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.054587 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.066534 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.079342 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.093106 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.106102 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.117495 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.122080 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.122127 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.122139 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.122159 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.122171 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:58Z","lastTransitionTime":"2025-11-25T13:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.132313 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.146746 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.160963 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.175593 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.190932 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.206499 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.215788 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.228463 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.228523 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.228539 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.228559 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.228571 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:58Z","lastTransitionTime":"2025-11-25T13:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.228576 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.250120 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.263553 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.276557 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.288434 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.299607 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.315912 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.326575 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.331524 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.331736 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.331830 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.331925 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.332012 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:58Z","lastTransitionTime":"2025-11-25T13:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.334913 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9e735a90-cf24-47de-b0e6-57d27a971b21-cni-binary-copy\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.334964 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b142b34b-79ab-4995-9578-d8fdf04e1fba-cni-binary-copy\") pod \"multus-additional-cni-plugins-wcs7x\" (UID: \"b142b34b-79ab-4995-9578-d8fdf04e1fba\") " pod="openshift-multus/multus-additional-cni-plugins-wcs7x" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.334991 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/b142b34b-79ab-4995-9578-d8fdf04e1fba-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-wcs7x\" (UID: \"b142b34b-79ab-4995-9578-d8fdf04e1fba\") " pod="openshift-multus/multus-additional-cni-plugins-wcs7x" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.335032 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/9e735a90-cf24-47de-b0e6-57d27a971b21-multus-daemon-config\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.336054 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/b142b34b-79ab-4995-9578-d8fdf04e1fba-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-wcs7x\" (UID: \"b142b34b-79ab-4995-9578-d8fdf04e1fba\") " pod="openshift-multus/multus-additional-cni-plugins-wcs7x" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.336147 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b142b34b-79ab-4995-9578-d8fdf04e1fba-cni-binary-copy\") pod \"multus-additional-cni-plugins-wcs7x\" (UID: \"b142b34b-79ab-4995-9578-d8fdf04e1fba\") " pod="openshift-multus/multus-additional-cni-plugins-wcs7x" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.336176 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/9e735a90-cf24-47de-b0e6-57d27a971b21-multus-daemon-config\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.336158 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9e735a90-cf24-47de-b0e6-57d27a971b21-cni-binary-copy\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.340547 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.354438 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.434603 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.434650 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.434661 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.434679 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.434691 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:58Z","lastTransitionTime":"2025-11-25T13:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.435605 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.435748 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:46:58 crc kubenswrapper[4702]: E1125 13:46:58.435778 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:47:06.435754275 +0000 UTC m=+35.598404725 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.435812 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:46:58 crc kubenswrapper[4702]: E1125 13:46:58.435867 4702 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 13:46:58 crc kubenswrapper[4702]: E1125 13:46:58.435912 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 13:47:06.435901178 +0000 UTC m=+35.598551628 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 13:46:58 crc kubenswrapper[4702]: E1125 13:46:58.435966 4702 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 13:46:58 crc kubenswrapper[4702]: E1125 13:46:58.436007 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 13:47:06.435998131 +0000 UTC m=+35.598648581 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.536448 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sj98\" (UniqueName: \"kubernetes.io/projected/9e735a90-cf24-47de-b0e6-57d27a971b21-kube-api-access-6sj98\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.536836 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.536873 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmrxm\" (UniqueName: \"kubernetes.io/projected/b142b34b-79ab-4995-9578-d8fdf04e1fba-kube-api-access-fmrxm\") pod \"multus-additional-cni-plugins-wcs7x\" (UID: \"b142b34b-79ab-4995-9578-d8fdf04e1fba\") " pod="openshift-multus/multus-additional-cni-plugins-wcs7x" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.536900 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:46:58 crc kubenswrapper[4702]: E1125 13:46:58.536932 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 13:46:58 crc kubenswrapper[4702]: E1125 13:46:58.536968 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 13:46:58 crc kubenswrapper[4702]: E1125 13:46:58.536986 4702 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 13:46:58 crc kubenswrapper[4702]: E1125 13:46:58.537046 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 13:46:58 crc kubenswrapper[4702]: E1125 13:46:58.537061 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 13:47:06.537037956 +0000 UTC m=+35.699688466 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 13:46:58 crc kubenswrapper[4702]: E1125 13:46:58.537065 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 13:46:58 crc kubenswrapper[4702]: E1125 13:46:58.537085 4702 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 13:46:58 crc kubenswrapper[4702]: E1125 13:46:58.537150 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 13:47:06.537130119 +0000 UTC m=+35.699780649 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.543747 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sj98\" (UniqueName: \"kubernetes.io/projected/9e735a90-cf24-47de-b0e6-57d27a971b21-kube-api-access-6sj98\") pod \"multus-gd2xs\" (UID: \"9e735a90-cf24-47de-b0e6-57d27a971b21\") " pod="openshift-multus/multus-gd2xs" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.544870 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.544940 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.544954 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.544974 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.544988 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:58Z","lastTransitionTime":"2025-11-25T13:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.546064 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmrxm\" (UniqueName: \"kubernetes.io/projected/b142b34b-79ab-4995-9578-d8fdf04e1fba-kube-api-access-fmrxm\") pod \"multus-additional-cni-plugins-wcs7x\" (UID: \"b142b34b-79ab-4995-9578-d8fdf04e1fba\") " pod="openshift-multus/multus-additional-cni-plugins-wcs7x" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.647425 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.647468 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.647481 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.647496 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.647524 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:58Z","lastTransitionTime":"2025-11-25T13:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.750834 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.751123 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.752322 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.752416 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.752556 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:58Z","lastTransitionTime":"2025-11-25T13:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.784632 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-gd2xs" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.790944 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.800843 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:46:58 crc kubenswrapper[4702]: E1125 13:46:58.801248 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.800969 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:46:58 crc kubenswrapper[4702]: E1125 13:46:58.801330 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.800943 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:46:58 crc kubenswrapper[4702]: E1125 13:46:58.801387 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:46:58 crc kubenswrapper[4702]: W1125 13:46:58.808717 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb142b34b_79ab_4995_9578_d8fdf04e1fba.slice/crio-86e66c42faa2ce91e5db7a8a0939eb88cd844c8e123cf1ad3fa081b6054a38e0 WatchSource:0}: Error finding container 86e66c42faa2ce91e5db7a8a0939eb88cd844c8e123cf1ad3fa081b6054a38e0: Status 404 returned error can't find the container with id 86e66c42faa2ce91e5db7a8a0939eb88cd844c8e123cf1ad3fa081b6054a38e0 Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.856001 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.856109 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.856117 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.856131 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.856142 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:58Z","lastTransitionTime":"2025-11-25T13:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.954027 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gd2xs" event={"ID":"9e735a90-cf24-47de-b0e6-57d27a971b21","Type":"ContainerStarted","Data":"6e974bd98038140376577175084fecede1544e3d0300b09cb23f77e8b7a6d77f"} Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.959083 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.959118 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.959126 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.959142 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.959152 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:58Z","lastTransitionTime":"2025-11-25T13:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.965316 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" event={"ID":"3c8a3234-01d3-48a9-a5b4-acb46b4218dc","Type":"ContainerStarted","Data":"7b2358ac7b25088c8d1b1cc5648a1fcbfd15211826d4c47f2c98839b9924138b"} Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.965366 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" event={"ID":"3c8a3234-01d3-48a9-a5b4-acb46b4218dc","Type":"ContainerStarted","Data":"a6d1bfac94a5fe6b382a9a3783250d6758603d00db9ce12ea44a3b54e6018103"} Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.965376 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" event={"ID":"3c8a3234-01d3-48a9-a5b4-acb46b4218dc","Type":"ContainerStarted","Data":"8a8fa25e9591d484f8dcb56974242924d4f31c7aef86b4376092842e57e1653c"} Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.965385 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" event={"ID":"3c8a3234-01d3-48a9-a5b4-acb46b4218dc","Type":"ContainerStarted","Data":"9649c45bc796d60538bc1edc8f1239c7c5445de7b2f3dbf57dbe7df6ecab7d6c"} Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.965394 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" event={"ID":"3c8a3234-01d3-48a9-a5b4-acb46b4218dc","Type":"ContainerStarted","Data":"a5dfe629efca9398d6dcc541b5315c5b6d8e036cba1c94064883785b2cc7a1a4"} Nov 25 13:46:58 crc kubenswrapper[4702]: I1125 13:46:58.968927 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" event={"ID":"b142b34b-79ab-4995-9578-d8fdf04e1fba","Type":"ContainerStarted","Data":"86e66c42faa2ce91e5db7a8a0939eb88cd844c8e123cf1ad3fa081b6054a38e0"} Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.062014 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.062519 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.062534 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.062551 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.062564 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:59Z","lastTransitionTime":"2025-11-25T13:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.165608 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.165645 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.165653 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.165670 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.165681 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:59Z","lastTransitionTime":"2025-11-25T13:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.268083 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.268136 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.268150 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.268171 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.268185 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:59Z","lastTransitionTime":"2025-11-25T13:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.370790 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.370833 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.370844 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.370861 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.370873 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:59Z","lastTransitionTime":"2025-11-25T13:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.473858 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.473959 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.473974 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.473993 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.474004 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:59Z","lastTransitionTime":"2025-11-25T13:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.577877 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.577920 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.577931 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.577956 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.577969 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:59Z","lastTransitionTime":"2025-11-25T13:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.599416 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-bhpwl"] Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.599812 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-bhpwl" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.604093 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.608940 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.608989 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.609093 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.626296 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:59Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.648600 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:59Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.654463 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e607da52-3cea-4ebe-b9e4-4eef40d6e714-serviceca\") pod \"node-ca-bhpwl\" (UID: \"e607da52-3cea-4ebe-b9e4-4eef40d6e714\") " pod="openshift-image-registry/node-ca-bhpwl" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.654516 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e607da52-3cea-4ebe-b9e4-4eef40d6e714-host\") pod \"node-ca-bhpwl\" (UID: \"e607da52-3cea-4ebe-b9e4-4eef40d6e714\") " pod="openshift-image-registry/node-ca-bhpwl" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.654638 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6jwz\" (UniqueName: \"kubernetes.io/projected/e607da52-3cea-4ebe-b9e4-4eef40d6e714-kube-api-access-j6jwz\") pod \"node-ca-bhpwl\" (UID: \"e607da52-3cea-4ebe-b9e4-4eef40d6e714\") " pod="openshift-image-registry/node-ca-bhpwl" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.660784 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:59Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.674178 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:59Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.679721 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.679770 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.679782 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.679799 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.679812 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:59Z","lastTransitionTime":"2025-11-25T13:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.688985 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:59Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.700371 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:59Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.712046 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:59Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.724115 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:59Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.747574 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:59Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.756111 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6jwz\" (UniqueName: \"kubernetes.io/projected/e607da52-3cea-4ebe-b9e4-4eef40d6e714-kube-api-access-j6jwz\") pod \"node-ca-bhpwl\" (UID: \"e607da52-3cea-4ebe-b9e4-4eef40d6e714\") " pod="openshift-image-registry/node-ca-bhpwl" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.756152 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e607da52-3cea-4ebe-b9e4-4eef40d6e714-host\") pod \"node-ca-bhpwl\" (UID: \"e607da52-3cea-4ebe-b9e4-4eef40d6e714\") " pod="openshift-image-registry/node-ca-bhpwl" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.756174 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e607da52-3cea-4ebe-b9e4-4eef40d6e714-serviceca\") pod \"node-ca-bhpwl\" (UID: \"e607da52-3cea-4ebe-b9e4-4eef40d6e714\") " pod="openshift-image-registry/node-ca-bhpwl" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.756309 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e607da52-3cea-4ebe-b9e4-4eef40d6e714-host\") pod \"node-ca-bhpwl\" (UID: \"e607da52-3cea-4ebe-b9e4-4eef40d6e714\") " pod="openshift-image-registry/node-ca-bhpwl" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.757202 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e607da52-3cea-4ebe-b9e4-4eef40d6e714-serviceca\") pod \"node-ca-bhpwl\" (UID: \"e607da52-3cea-4ebe-b9e4-4eef40d6e714\") " pod="openshift-image-registry/node-ca-bhpwl" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.761440 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:59Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.777496 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:59Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.778198 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6jwz\" (UniqueName: \"kubernetes.io/projected/e607da52-3cea-4ebe-b9e4-4eef40d6e714-kube-api-access-j6jwz\") pod \"node-ca-bhpwl\" (UID: \"e607da52-3cea-4ebe-b9e4-4eef40d6e714\") " pod="openshift-image-registry/node-ca-bhpwl" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.781708 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.781744 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.781754 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.781796 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.781809 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:59Z","lastTransitionTime":"2025-11-25T13:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.792963 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:59Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.807807 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:59Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.827693 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:59Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.842069 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:46:59Z is after 2025-08-24T17:21:41Z" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.884277 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.884336 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.884347 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.884429 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.884444 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:59Z","lastTransitionTime":"2025-11-25T13:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.912979 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-bhpwl" Nov 25 13:46:59 crc kubenswrapper[4702]: W1125 13:46:59.925424 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode607da52_3cea_4ebe_b9e4_4eef40d6e714.slice/crio-fcd6e8315e9553a3283d2f240123d81197e38b9d71385b49faf1368c281f582a WatchSource:0}: Error finding container fcd6e8315e9553a3283d2f240123d81197e38b9d71385b49faf1368c281f582a: Status 404 returned error can't find the container with id fcd6e8315e9553a3283d2f240123d81197e38b9d71385b49faf1368c281f582a Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.972306 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gd2xs" event={"ID":"9e735a90-cf24-47de-b0e6-57d27a971b21","Type":"ContainerStarted","Data":"b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694"} Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.974748 4702 generic.go:334] "Generic (PLEG): container finished" podID="b142b34b-79ab-4995-9578-d8fdf04e1fba" containerID="76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711" exitCode=0 Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.974800 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" event={"ID":"b142b34b-79ab-4995-9578-d8fdf04e1fba","Type":"ContainerDied","Data":"76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711"} Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.977605 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-bhpwl" event={"ID":"e607da52-3cea-4ebe-b9e4-4eef40d6e714","Type":"ContainerStarted","Data":"fcd6e8315e9553a3283d2f240123d81197e38b9d71385b49faf1368c281f582a"} Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.986682 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.986765 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.986775 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.986792 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.986806 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:46:59Z","lastTransitionTime":"2025-11-25T13:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:46:59 crc kubenswrapper[4702]: I1125 13:46:59.987327 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" event={"ID":"3c8a3234-01d3-48a9-a5b4-acb46b4218dc","Type":"ContainerStarted","Data":"c91a144aef48f19006041105594111c382420f9e6556a64580f6952990ef4db4"} Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.005273 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.023465 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.039571 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.052123 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.069892 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.088647 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.091516 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.091539 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.091546 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.091558 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.091567 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:00Z","lastTransitionTime":"2025-11-25T13:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.103282 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.118487 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.131606 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.151273 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.168950 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.182319 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.194928 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.195035 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.195112 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.195127 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.195146 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.195159 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:00Z","lastTransitionTime":"2025-11-25T13:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.214900 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.232184 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.253254 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.269125 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.282931 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.292826 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.302026 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.302064 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.302072 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.302086 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.302095 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:00Z","lastTransitionTime":"2025-11-25T13:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.311737 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.324025 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.337893 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.351326 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.362940 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.378860 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.391459 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.404343 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.404670 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.404715 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.404727 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.404740 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.404749 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:00Z","lastTransitionTime":"2025-11-25T13:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.415849 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.428145 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.441400 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.507668 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.507717 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.507732 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.507749 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.507764 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:00Z","lastTransitionTime":"2025-11-25T13:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.610038 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.610086 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.610101 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.610118 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.610132 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:00Z","lastTransitionTime":"2025-11-25T13:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.712663 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.712704 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.712725 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.712741 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.712756 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:00Z","lastTransitionTime":"2025-11-25T13:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.800737 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.800766 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.800831 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:47:00 crc kubenswrapper[4702]: E1125 13:47:00.800866 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:47:00 crc kubenswrapper[4702]: E1125 13:47:00.801057 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:47:00 crc kubenswrapper[4702]: E1125 13:47:00.801190 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.815008 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.815045 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.815054 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.815073 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.815091 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:00Z","lastTransitionTime":"2025-11-25T13:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.917789 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.917828 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.917848 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.917864 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.917875 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:00Z","lastTransitionTime":"2025-11-25T13:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.992048 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-bhpwl" event={"ID":"e607da52-3cea-4ebe-b9e4-4eef40d6e714","Type":"ContainerStarted","Data":"7d2efcd45f14690e23b56c19e3282342933f83acd4940f89dd4046b61e4a608b"} Nov 25 13:47:00 crc kubenswrapper[4702]: I1125 13:47:00.993613 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" event={"ID":"b142b34b-79ab-4995-9578-d8fdf04e1fba","Type":"ContainerStarted","Data":"15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8"} Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.006659 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.018709 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.019827 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.019858 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.019867 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.019883 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.019895 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:01Z","lastTransitionTime":"2025-11-25T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.029480 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.047343 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.061991 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.076548 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.087630 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.099328 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.112762 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.122669 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.122725 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.122738 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.122772 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.122789 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:01Z","lastTransitionTime":"2025-11-25T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.134926 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.152909 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.168005 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.182500 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.194571 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.207427 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2efcd45f14690e23b56c19e3282342933f83acd4940f89dd4046b61e4a608b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.220168 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.225185 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.225253 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.225269 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.225292 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.225304 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:01Z","lastTransitionTime":"2025-11-25T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.235666 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.257380 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.270461 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.287891 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.299369 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.314405 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.325610 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2efcd45f14690e23b56c19e3282342933f83acd4940f89dd4046b61e4a608b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.327707 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.327745 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.327755 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.327769 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.327779 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:01Z","lastTransitionTime":"2025-11-25T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.340152 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.354036 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.366737 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.387742 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.404772 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.419623 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.430946 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.430988 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.430999 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.431013 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.431025 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:01Z","lastTransitionTime":"2025-11-25T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.433287 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.533142 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.533295 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.533310 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.533326 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.533338 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:01Z","lastTransitionTime":"2025-11-25T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.635757 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.635798 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.635807 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.635821 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.635830 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:01Z","lastTransitionTime":"2025-11-25T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.737836 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.737876 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.737886 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.737901 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.737913 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:01Z","lastTransitionTime":"2025-11-25T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.814274 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.831901 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.839988 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.840031 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.840041 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.840057 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.840069 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:01Z","lastTransitionTime":"2025-11-25T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.857339 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.871632 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.885293 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.895505 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.906786 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.918337 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2efcd45f14690e23b56c19e3282342933f83acd4940f89dd4046b61e4a608b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.930469 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.942571 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.942621 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.942637 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.942653 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.942666 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:01Z","lastTransitionTime":"2025-11-25T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.944942 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.956426 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.975147 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:01 crc kubenswrapper[4702]: I1125 13:47:01.987528 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.001968 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.002146 4702 generic.go:334] "Generic (PLEG): container finished" podID="b142b34b-79ab-4995-9578-d8fdf04e1fba" containerID="15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8" exitCode=0 Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.002226 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" event={"ID":"b142b34b-79ab-4995-9578-d8fdf04e1fba","Type":"ContainerDied","Data":"15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8"} Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.008940 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" event={"ID":"3c8a3234-01d3-48a9-a5b4-acb46b4218dc","Type":"ContainerStarted","Data":"8700d12c2e236e704fc60136a51489d1d6890a83621480c23a6168d4a0f18f46"} Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.014132 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.028880 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.045018 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.045053 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.045066 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.045083 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.045096 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:02Z","lastTransitionTime":"2025-11-25T13:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.046218 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.063455 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.073762 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.090680 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.103680 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2efcd45f14690e23b56c19e3282342933f83acd4940f89dd4046b61e4a608b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.128415 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.144190 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.148864 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.148960 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.148979 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.149056 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.149081 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:02Z","lastTransitionTime":"2025-11-25T13:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.160061 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.175806 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.191479 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.215341 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.229435 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.245662 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.252010 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.252047 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.252060 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.252077 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.252089 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:02Z","lastTransitionTime":"2025-11-25T13:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.260549 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.354616 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.354758 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.354785 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.354856 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.354880 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:02Z","lastTransitionTime":"2025-11-25T13:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.458144 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.458195 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.458207 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.458228 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.458264 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:02Z","lastTransitionTime":"2025-11-25T13:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.561358 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.561909 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.562013 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.562131 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.562207 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:02Z","lastTransitionTime":"2025-11-25T13:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.665472 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.666014 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.666032 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.666056 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.666075 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:02Z","lastTransitionTime":"2025-11-25T13:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.769600 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.769652 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.769663 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.769683 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.769696 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:02Z","lastTransitionTime":"2025-11-25T13:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.800313 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.800432 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.800556 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:02 crc kubenswrapper[4702]: E1125 13:47:02.800659 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:47:02 crc kubenswrapper[4702]: E1125 13:47:02.800689 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:47:02 crc kubenswrapper[4702]: E1125 13:47:02.800753 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.873115 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.873189 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.873206 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.873252 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.873278 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:02Z","lastTransitionTime":"2025-11-25T13:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.975625 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.975666 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.975677 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.975692 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:02 crc kubenswrapper[4702]: I1125 13:47:02.975703 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:02Z","lastTransitionTime":"2025-11-25T13:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.015856 4702 generic.go:334] "Generic (PLEG): container finished" podID="b142b34b-79ab-4995-9578-d8fdf04e1fba" containerID="7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a" exitCode=0 Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.015906 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" event={"ID":"b142b34b-79ab-4995-9578-d8fdf04e1fba","Type":"ContainerDied","Data":"7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a"} Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.034447 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2efcd45f14690e23b56c19e3282342933f83acd4940f89dd4046b61e4a608b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:03Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.057435 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:03Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.074538 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:03Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.080157 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.080209 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.080220 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.080257 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.080271 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:03Z","lastTransitionTime":"2025-11-25T13:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.093967 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:03Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.105629 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:03Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.118863 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:03Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.138922 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:03Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.156198 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:03Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.195717 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:03Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.214256 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.214326 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.214343 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.214363 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.214381 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:03Z","lastTransitionTime":"2025-11-25T13:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.238712 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:03Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.259455 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:03Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.273914 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:03Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.287848 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:03Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.303913 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:03Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.317351 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.317428 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.317442 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.317467 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.317485 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:03Z","lastTransitionTime":"2025-11-25T13:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.322309 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:03Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.419859 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.419925 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.419935 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.419959 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.419973 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:03Z","lastTransitionTime":"2025-11-25T13:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.522360 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.522404 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.522417 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.522432 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.522442 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:03Z","lastTransitionTime":"2025-11-25T13:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.624951 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.625013 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.625032 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.625051 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.625066 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:03Z","lastTransitionTime":"2025-11-25T13:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.727535 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.727571 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.727580 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.727595 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.727605 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:03Z","lastTransitionTime":"2025-11-25T13:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.829698 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.829735 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.829744 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.829759 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.829768 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:03Z","lastTransitionTime":"2025-11-25T13:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.931565 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.931615 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.931625 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.931645 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:03 crc kubenswrapper[4702]: I1125 13:47:03.931662 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:03Z","lastTransitionTime":"2025-11-25T13:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.028879 4702 generic.go:334] "Generic (PLEG): container finished" podID="b142b34b-79ab-4995-9578-d8fdf04e1fba" containerID="06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6" exitCode=0 Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.028947 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" event={"ID":"b142b34b-79ab-4995-9578-d8fdf04e1fba","Type":"ContainerDied","Data":"06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6"} Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.034116 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.034197 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.034242 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.034292 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.034306 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:04Z","lastTransitionTime":"2025-11-25T13:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.041218 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" event={"ID":"3c8a3234-01d3-48a9-a5b4-acb46b4218dc","Type":"ContainerStarted","Data":"06049a90425ba8ed3b9fff368fb341653f3ba688dee562903b75cb65faeaaf51"} Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.041529 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.044121 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.059459 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.076884 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.097855 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.099336 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.116777 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.131027 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.136837 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.136910 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.136925 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.136944 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.136956 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:04Z","lastTransitionTime":"2025-11-25T13:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.148315 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.164655 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.182375 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.200177 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.212288 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.228525 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.239249 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2efcd45f14690e23b56c19e3282342933f83acd4940f89dd4046b61e4a608b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.240369 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.240398 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.240407 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.240425 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.240437 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:04Z","lastTransitionTime":"2025-11-25T13:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.264387 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.279668 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.295759 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.315486 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.325907 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2efcd45f14690e23b56c19e3282342933f83acd4940f89dd4046b61e4a608b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.344837 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.344909 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.344934 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.344968 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.345035 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:04Z","lastTransitionTime":"2025-11-25T13:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.348081 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.363665 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.376731 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.389425 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.404940 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.424840 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a8fa25e9591d484f8dcb56974242924d4f31c7aef86b4376092842e57e1653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d1bfac94a5fe6b382a9a3783250d6758603d00db9ce12ea44a3b54e6018103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c91a144aef48f19006041105594111c382420f9e6556a64580f6952990ef4db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2358ac7b25088c8d1b1cc5648a1fcbfd15211826d4c47f2c98839b9924138b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9649c45bc796d60538bc1edc8f1239c7c5445de7b2f3dbf57dbe7df6ecab7d6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5dfe629efca9398d6dcc541b5315c5b6d8e036cba1c94064883785b2cc7a1a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06049a90425ba8ed3b9fff368fb341653f3ba688dee562903b75cb65faeaaf51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8700d12c2e236e704fc60136a51489d1d6890a83621480c23a6168d4a0f18f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.438505 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.448278 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.448508 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.448575 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.448642 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.448709 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:04Z","lastTransitionTime":"2025-11-25T13:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.455023 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.469080 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.485266 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.500721 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.517578 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.552124 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.552200 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.552210 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.552247 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.552260 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:04Z","lastTransitionTime":"2025-11-25T13:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.655726 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.655779 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.655791 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.655810 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.655823 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:04Z","lastTransitionTime":"2025-11-25T13:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.759096 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.759202 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.759221 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.759478 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.759500 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:04Z","lastTransitionTime":"2025-11-25T13:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.800762 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.800957 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:04 crc kubenswrapper[4702]: E1125 13:47:04.800997 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:47:04 crc kubenswrapper[4702]: E1125 13:47:04.801339 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.801375 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:47:04 crc kubenswrapper[4702]: E1125 13:47:04.801559 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.862997 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.863070 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.863088 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.863117 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.863134 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:04Z","lastTransitionTime":"2025-11-25T13:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.965731 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.965782 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.965794 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.965813 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:04 crc kubenswrapper[4702]: I1125 13:47:04.965827 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:04Z","lastTransitionTime":"2025-11-25T13:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.012343 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.025914 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.042677 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.044601 4702 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.045256 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.063561 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a8fa25e9591d484f8dcb56974242924d4f31c7aef86b4376092842e57e1653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d1bfac94a5fe6b382a9a3783250d6758603d00db9ce12ea44a3b54e6018103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c91a144aef48f19006041105594111c382420f9e6556a64580f6952990ef4db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2358ac7b25088c8d1b1cc5648a1fcbfd15211826d4c47f2c98839b9924138b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9649c45bc796d60538bc1edc8f1239c7c5445de7b2f3dbf57dbe7df6ecab7d6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5dfe629efca9398d6dcc541b5315c5b6d8e036cba1c94064883785b2cc7a1a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06049a90425ba8ed3b9fff368fb341653f3ba688dee562903b75cb65faeaaf51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8700d12c2e236e704fc60136a51489d1d6890a83621480c23a6168d4a0f18f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.078194 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.093368 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.102284 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.102371 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.102384 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.102407 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.102424 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:05Z","lastTransitionTime":"2025-11-25T13:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.106285 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.109493 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.126542 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.145272 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.160400 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.175217 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.197037 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.205980 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.206026 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.206035 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.206054 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.206065 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:05Z","lastTransitionTime":"2025-11-25T13:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.213481 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2efcd45f14690e23b56c19e3282342933f83acd4940f89dd4046b61e4a608b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.233120 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.246365 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.256720 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.270523 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.282198 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2efcd45f14690e23b56c19e3282342933f83acd4940f89dd4046b61e4a608b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.303304 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.308781 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.308843 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.308853 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.308880 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.308893 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:05Z","lastTransitionTime":"2025-11-25T13:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.317346 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.332006 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.342393 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.354793 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.374618 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a8fa25e9591d484f8dcb56974242924d4f31c7aef86b4376092842e57e1653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d1bfac94a5fe6b382a9a3783250d6758603d00db9ce12ea44a3b54e6018103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c91a144aef48f19006041105594111c382420f9e6556a64580f6952990ef4db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2358ac7b25088c8d1b1cc5648a1fcbfd15211826d4c47f2c98839b9924138b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9649c45bc796d60538bc1edc8f1239c7c5445de7b2f3dbf57dbe7df6ecab7d6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5dfe629efca9398d6dcc541b5315c5b6d8e036cba1c94064883785b2cc7a1a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06049a90425ba8ed3b9fff368fb341653f3ba688dee562903b75cb65faeaaf51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8700d12c2e236e704fc60136a51489d1d6890a83621480c23a6168d4a0f18f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.388927 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.400997 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.411688 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.411736 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.411747 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.411767 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.411779 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:05Z","lastTransitionTime":"2025-11-25T13:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.412812 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.425697 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.440848 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.452994 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.467712 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.513653 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.513683 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.513692 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.513706 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.513716 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:05Z","lastTransitionTime":"2025-11-25T13:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.615511 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.615562 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.615571 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.615583 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.615591 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:05Z","lastTransitionTime":"2025-11-25T13:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.717274 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.717514 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.717671 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.717773 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.717850 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:05Z","lastTransitionTime":"2025-11-25T13:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.820266 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.820304 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.820317 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.820334 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.820349 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:05Z","lastTransitionTime":"2025-11-25T13:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.922933 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.922988 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.922999 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.923015 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:05 crc kubenswrapper[4702]: I1125 13:47:05.923027 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:05Z","lastTransitionTime":"2025-11-25T13:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.025744 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.025808 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.025822 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.025849 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.025894 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:06Z","lastTransitionTime":"2025-11-25T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.050628 4702 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.050643 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" event={"ID":"b142b34b-79ab-4995-9578-d8fdf04e1fba","Type":"ContainerStarted","Data":"b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9"} Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.129169 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.129228 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.129253 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.129270 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.129280 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:06Z","lastTransitionTime":"2025-11-25T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.232678 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.232709 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.232718 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.232734 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.232744 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:06Z","lastTransitionTime":"2025-11-25T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.334533 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.334582 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.334594 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.334613 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.334630 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:06Z","lastTransitionTime":"2025-11-25T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.437601 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.437649 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.437660 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.437677 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.437688 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:06Z","lastTransitionTime":"2025-11-25T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.535378 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.535659 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.535748 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:06 crc kubenswrapper[4702]: E1125 13:47:06.535782 4702 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 13:47:06 crc kubenswrapper[4702]: E1125 13:47:06.535794 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:47:22.535764906 +0000 UTC m=+51.698415356 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:47:06 crc kubenswrapper[4702]: E1125 13:47:06.535930 4702 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 13:47:06 crc kubenswrapper[4702]: E1125 13:47:06.536089 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 13:47:22.536080094 +0000 UTC m=+51.698730544 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 13:47:06 crc kubenswrapper[4702]: E1125 13:47:06.536146 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 13:47:22.536111314 +0000 UTC m=+51.698761944 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.541615 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.541659 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.541669 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.541685 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.541696 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:06Z","lastTransitionTime":"2025-11-25T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.637842 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.637963 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:47:06 crc kubenswrapper[4702]: E1125 13:47:06.638325 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 13:47:06 crc kubenswrapper[4702]: E1125 13:47:06.638363 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 13:47:06 crc kubenswrapper[4702]: E1125 13:47:06.638387 4702 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 13:47:06 crc kubenswrapper[4702]: E1125 13:47:06.638478 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 13:47:22.638447512 +0000 UTC m=+51.801098002 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 13:47:06 crc kubenswrapper[4702]: E1125 13:47:06.638685 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 13:47:06 crc kubenswrapper[4702]: E1125 13:47:06.638749 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 13:47:06 crc kubenswrapper[4702]: E1125 13:47:06.638777 4702 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 13:47:06 crc kubenswrapper[4702]: E1125 13:47:06.638892 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 13:47:22.638862502 +0000 UTC m=+51.801512962 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.645427 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.645480 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.645500 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.645529 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.645549 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:06Z","lastTransitionTime":"2025-11-25T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.748002 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.748054 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.748069 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.748088 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.748102 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:06Z","lastTransitionTime":"2025-11-25T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.800815 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:47:06 crc kubenswrapper[4702]: E1125 13:47:06.801267 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.800977 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:06 crc kubenswrapper[4702]: E1125 13:47:06.801520 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.800915 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:47:06 crc kubenswrapper[4702]: E1125 13:47:06.801726 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.851448 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.851557 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.851583 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.851624 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.851651 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:06Z","lastTransitionTime":"2025-11-25T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.954864 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.954938 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.954954 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.954972 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:06 crc kubenswrapper[4702]: I1125 13:47:06.954983 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:06Z","lastTransitionTime":"2025-11-25T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.054647 4702 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.057394 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.057426 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.057438 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.057455 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.057468 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:07Z","lastTransitionTime":"2025-11-25T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.069126 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:07Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.089310 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:07Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.110645 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:07Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.125470 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2efcd45f14690e23b56c19e3282342933f83acd4940f89dd4046b61e4a608b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:07Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.147836 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:07Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.161155 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.161387 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.161505 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.161601 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.161706 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:07Z","lastTransitionTime":"2025-11-25T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.165396 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:07Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.184656 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:07Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.201035 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:07Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.218184 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.218273 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.218297 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.218324 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.218346 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:07Z","lastTransitionTime":"2025-11-25T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.218685 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:07Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:07 crc kubenswrapper[4702]: E1125 13:47:07.237312 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"96cc93a3-66f3-477f-bbc8-ef3a984ed406\\\",\\\"systemUUID\\\":\\\"bb366f18-cc3c-451f-9d6a-c0cc41864b99\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:07Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.245165 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.245122 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a8fa25e9591d484f8dcb56974242924d4f31c7aef86b4376092842e57e1653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d1bfac94a5fe6b382a9a3783250d6758603d00db9ce12ea44a3b54e6018103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c91a144aef48f19006041105594111c382420f9e6556a64580f6952990ef4db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2358ac7b25088c8d1b1cc5648a1fcbfd15211826d4c47f2c98839b9924138b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9649c45bc796d60538bc1edc8f1239c7c5445de7b2f3dbf57dbe7df6ecab7d6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5dfe629efca9398d6dcc541b5315c5b6d8e036cba1c94064883785b2cc7a1a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06049a90425ba8ed3b9fff368fb341653f3ba688dee562903b75cb65faeaaf51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8700d12c2e236e704fc60136a51489d1d6890a83621480c23a6168d4a0f18f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:07Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.245256 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.245276 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.245296 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.245314 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:07Z","lastTransitionTime":"2025-11-25T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:07 crc kubenswrapper[4702]: E1125 13:47:07.261776 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"96cc93a3-66f3-477f-bbc8-ef3a984ed406\\\",\\\"systemUUID\\\":\\\"bb366f18-cc3c-451f-9d6a-c0cc41864b99\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:07Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.265393 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:07Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.266562 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.266604 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.266614 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.266632 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.266642 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:07Z","lastTransitionTime":"2025-11-25T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.279912 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:07Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:07 crc kubenswrapper[4702]: E1125 13:47:07.283464 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"96cc93a3-66f3-477f-bbc8-ef3a984ed406\\\",\\\"systemUUID\\\":\\\"bb366f18-cc3c-451f-9d6a-c0cc41864b99\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:07Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.288355 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.288412 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.288426 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.288448 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.288463 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:07Z","lastTransitionTime":"2025-11-25T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.298399 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:07Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:07 crc kubenswrapper[4702]: E1125 13:47:07.303798 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"96cc93a3-66f3-477f-bbc8-ef3a984ed406\\\",\\\"systemUUID\\\":\\\"bb366f18-cc3c-451f-9d6a-c0cc41864b99\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:07Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.308088 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.308121 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.308133 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.308152 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.308165 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:07Z","lastTransitionTime":"2025-11-25T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.314857 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:07Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:07 crc kubenswrapper[4702]: E1125 13:47:07.324186 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"96cc93a3-66f3-477f-bbc8-ef3a984ed406\\\",\\\"systemUUID\\\":\\\"bb366f18-cc3c-451f-9d6a-c0cc41864b99\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:07Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:07 crc kubenswrapper[4702]: E1125 13:47:07.324683 4702 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.326800 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.326926 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.327338 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.327458 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.327520 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:07Z","lastTransitionTime":"2025-11-25T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.329654 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:07Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.430459 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.430489 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.430501 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.430519 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.430530 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:07Z","lastTransitionTime":"2025-11-25T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.533646 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.533692 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.533704 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.533721 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.533734 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:07Z","lastTransitionTime":"2025-11-25T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.636582 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.636934 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.637056 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.637149 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.637282 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:07Z","lastTransitionTime":"2025-11-25T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.741303 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.741370 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.741388 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.741416 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.741438 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:07Z","lastTransitionTime":"2025-11-25T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.843729 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.843811 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.843829 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.843860 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.843881 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:07Z","lastTransitionTime":"2025-11-25T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.947288 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.947354 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.947367 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.947393 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:07 crc kubenswrapper[4702]: I1125 13:47:07.947410 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:07Z","lastTransitionTime":"2025-11-25T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.051413 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.051475 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.051489 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.051511 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.051526 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:08Z","lastTransitionTime":"2025-11-25T13:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.063700 4702 generic.go:334] "Generic (PLEG): container finished" podID="b142b34b-79ab-4995-9578-d8fdf04e1fba" containerID="b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9" exitCode=0 Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.064529 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" event={"ID":"b142b34b-79ab-4995-9578-d8fdf04e1fba","Type":"ContainerDied","Data":"b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9"} Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.092433 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:08Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.114065 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:08Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.136725 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:08Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.154596 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.154635 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.154645 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.154663 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.154676 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:08Z","lastTransitionTime":"2025-11-25T13:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.161925 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:08Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.192017 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:08Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.219921 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:08Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.246751 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:08Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.258095 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.258154 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.258169 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.258198 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.258217 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:08Z","lastTransitionTime":"2025-11-25T13:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.268650 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:08Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.284968 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:08Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.302663 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:08Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.320625 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2efcd45f14690e23b56c19e3282342933f83acd4940f89dd4046b61e4a608b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:08Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.336707 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:08Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.352648 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:08Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.360848 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.361033 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.361164 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.361321 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.361443 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:08Z","lastTransitionTime":"2025-11-25T13:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.367023 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:08Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.387028 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a8fa25e9591d484f8dcb56974242924d4f31c7aef86b4376092842e57e1653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d1bfac94a5fe6b382a9a3783250d6758603d00db9ce12ea44a3b54e6018103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c91a144aef48f19006041105594111c382420f9e6556a64580f6952990ef4db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2358ac7b25088c8d1b1cc5648a1fcbfd15211826d4c47f2c98839b9924138b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9649c45bc796d60538bc1edc8f1239c7c5445de7b2f3dbf57dbe7df6ecab7d6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5dfe629efca9398d6dcc541b5315c5b6d8e036cba1c94064883785b2cc7a1a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06049a90425ba8ed3b9fff368fb341653f3ba688dee562903b75cb65faeaaf51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8700d12c2e236e704fc60136a51489d1d6890a83621480c23a6168d4a0f18f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:08Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.465064 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.465126 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.465144 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.465163 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.465208 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:08Z","lastTransitionTime":"2025-11-25T13:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.568506 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.568579 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.568599 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.568625 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.568648 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:08Z","lastTransitionTime":"2025-11-25T13:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.672168 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.672265 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.672284 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.672310 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.672330 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:08Z","lastTransitionTime":"2025-11-25T13:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.775353 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.775436 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.775454 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.775477 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.775494 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:08Z","lastTransitionTime":"2025-11-25T13:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.801003 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.801114 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:08 crc kubenswrapper[4702]: E1125 13:47:08.801300 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:47:08 crc kubenswrapper[4702]: E1125 13:47:08.801497 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.801651 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:47:08 crc kubenswrapper[4702]: E1125 13:47:08.801754 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.878672 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.878716 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.878734 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.878758 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.878776 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:08Z","lastTransitionTime":"2025-11-25T13:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.981893 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.981945 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.981978 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.982001 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:08 crc kubenswrapper[4702]: I1125 13:47:08.982018 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:08Z","lastTransitionTime":"2025-11-25T13:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.075886 4702 generic.go:334] "Generic (PLEG): container finished" podID="b142b34b-79ab-4995-9578-d8fdf04e1fba" containerID="6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662" exitCode=0 Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.075933 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" event={"ID":"b142b34b-79ab-4995-9578-d8fdf04e1fba","Type":"ContainerDied","Data":"6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662"} Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.084919 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.085092 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.085292 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.085417 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.085527 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:09Z","lastTransitionTime":"2025-11-25T13:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.115078 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:09Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.130883 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:09Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.156513 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:09Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.175676 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:09Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.188174 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.188209 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.188222 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.188259 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.188270 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:09Z","lastTransitionTime":"2025-11-25T13:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.192306 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:09Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.213999 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:09Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.246393 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:09Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.271372 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:09Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.284260 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:09Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.290510 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.290547 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.290558 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.290572 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.290582 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:09Z","lastTransitionTime":"2025-11-25T13:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.300150 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:09Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.315174 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2efcd45f14690e23b56c19e3282342933f83acd4940f89dd4046b61e4a608b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:09Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.328042 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:09Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.341002 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:09Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.353533 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:09Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.374768 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a8fa25e9591d484f8dcb56974242924d4f31c7aef86b4376092842e57e1653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d1bfac94a5fe6b382a9a3783250d6758603d00db9ce12ea44a3b54e6018103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c91a144aef48f19006041105594111c382420f9e6556a64580f6952990ef4db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2358ac7b25088c8d1b1cc5648a1fcbfd15211826d4c47f2c98839b9924138b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9649c45bc796d60538bc1edc8f1239c7c5445de7b2f3dbf57dbe7df6ecab7d6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5dfe629efca9398d6dcc541b5315c5b6d8e036cba1c94064883785b2cc7a1a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06049a90425ba8ed3b9fff368fb341653f3ba688dee562903b75cb65faeaaf51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8700d12c2e236e704fc60136a51489d1d6890a83621480c23a6168d4a0f18f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:09Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.393631 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.393669 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.393677 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.393695 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.393706 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:09Z","lastTransitionTime":"2025-11-25T13:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.496487 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.496533 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.496543 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.496560 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.496571 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:09Z","lastTransitionTime":"2025-11-25T13:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.599027 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.599091 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.599108 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.599130 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.599144 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:09Z","lastTransitionTime":"2025-11-25T13:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.702417 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.702453 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.702464 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.702478 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.702488 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:09Z","lastTransitionTime":"2025-11-25T13:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.796507 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv"] Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.796929 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.799618 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.809434 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.812331 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.812378 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.812391 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.812411 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.812428 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:09Z","lastTransitionTime":"2025-11-25T13:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.827615 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:09Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.864961 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:09Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.878414 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6jjm\" (UniqueName: \"kubernetes.io/projected/675f4b11-d943-4acc-afde-6f4f45585551-kube-api-access-z6jjm\") pod \"ovnkube-control-plane-749d76644c-djssv\" (UID: \"675f4b11-d943-4acc-afde-6f4f45585551\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.878487 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/675f4b11-d943-4acc-afde-6f4f45585551-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-djssv\" (UID: \"675f4b11-d943-4acc-afde-6f4f45585551\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.878557 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/675f4b11-d943-4acc-afde-6f4f45585551-env-overrides\") pod \"ovnkube-control-plane-749d76644c-djssv\" (UID: \"675f4b11-d943-4acc-afde-6f4f45585551\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.878679 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/675f4b11-d943-4acc-afde-6f4f45585551-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-djssv\" (UID: \"675f4b11-d943-4acc-afde-6f4f45585551\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.884899 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:09Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.900081 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:09Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.914948 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.914984 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.914997 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.915017 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.915033 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:09Z","lastTransitionTime":"2025-11-25T13:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.915782 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:09Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.940276 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:09Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.958841 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:09Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.975268 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:09Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.979541 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6jjm\" (UniqueName: \"kubernetes.io/projected/675f4b11-d943-4acc-afde-6f4f45585551-kube-api-access-z6jjm\") pod \"ovnkube-control-plane-749d76644c-djssv\" (UID: \"675f4b11-d943-4acc-afde-6f4f45585551\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.979607 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/675f4b11-d943-4acc-afde-6f4f45585551-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-djssv\" (UID: \"675f4b11-d943-4acc-afde-6f4f45585551\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.979633 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/675f4b11-d943-4acc-afde-6f4f45585551-env-overrides\") pod \"ovnkube-control-plane-749d76644c-djssv\" (UID: \"675f4b11-d943-4acc-afde-6f4f45585551\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.979687 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/675f4b11-d943-4acc-afde-6f4f45585551-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-djssv\" (UID: \"675f4b11-d943-4acc-afde-6f4f45585551\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.980433 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/675f4b11-d943-4acc-afde-6f4f45585551-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-djssv\" (UID: \"675f4b11-d943-4acc-afde-6f4f45585551\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.980483 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/675f4b11-d943-4acc-afde-6f4f45585551-env-overrides\") pod \"ovnkube-control-plane-749d76644c-djssv\" (UID: \"675f4b11-d943-4acc-afde-6f4f45585551\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.986106 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:09Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.987450 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/675f4b11-d943-4acc-afde-6f4f45585551-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-djssv\" (UID: \"675f4b11-d943-4acc-afde-6f4f45585551\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" Nov 25 13:47:09 crc kubenswrapper[4702]: I1125 13:47:09.996493 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6jjm\" (UniqueName: \"kubernetes.io/projected/675f4b11-d943-4acc-afde-6f4f45585551-kube-api-access-z6jjm\") pod \"ovnkube-control-plane-749d76644c-djssv\" (UID: \"675f4b11-d943-4acc-afde-6f4f45585551\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.000755 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:09Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.012804 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2efcd45f14690e23b56c19e3282342933f83acd4940f89dd4046b61e4a608b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.017074 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.017164 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.017178 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.017195 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.017207 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:10Z","lastTransitionTime":"2025-11-25T13:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.025088 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675f4b11-d943-4acc-afde-6f4f45585551\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-djssv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.039864 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.051035 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.060426 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.080553 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a8fa25e9591d484f8dcb56974242924d4f31c7aef86b4376092842e57e1653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d1bfac94a5fe6b382a9a3783250d6758603d00db9ce12ea44a3b54e6018103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c91a144aef48f19006041105594111c382420f9e6556a64580f6952990ef4db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2358ac7b25088c8d1b1cc5648a1fcbfd15211826d4c47f2c98839b9924138b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9649c45bc796d60538bc1edc8f1239c7c5445de7b2f3dbf57dbe7df6ecab7d6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5dfe629efca9398d6dcc541b5315c5b6d8e036cba1c94064883785b2cc7a1a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06049a90425ba8ed3b9fff368fb341653f3ba688dee562903b75cb65faeaaf51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8700d12c2e236e704fc60136a51489d1d6890a83621480c23a6168d4a0f18f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.083044 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" event={"ID":"b142b34b-79ab-4995-9578-d8fdf04e1fba","Type":"ContainerStarted","Data":"60a7f9a29a1e574b4eb0fc72705f8c96b6f613f8a7c164605070077212854cb1"} Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.096419 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.109480 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60a7f9a29a1e574b4eb0fc72705f8c96b6f613f8a7c164605070077212854cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.120585 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.120627 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.120638 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.120654 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.120665 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:10Z","lastTransitionTime":"2025-11-25T13:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.128865 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.129706 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: W1125 13:47:10.146613 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod675f4b11_d943_4acc_afde_6f4f45585551.slice/crio-62b27504ebe723ca2b16bb4fcb8c5611cffffc7a47cb356c072e2cc5aa18e796 WatchSource:0}: Error finding container 62b27504ebe723ca2b16bb4fcb8c5611cffffc7a47cb356c072e2cc5aa18e796: Status 404 returned error can't find the container with id 62b27504ebe723ca2b16bb4fcb8c5611cffffc7a47cb356c072e2cc5aa18e796 Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.147158 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.169292 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.184458 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.201174 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.214214 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2efcd45f14690e23b56c19e3282342933f83acd4940f89dd4046b61e4a608b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.222293 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.222326 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.222333 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.222347 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.222358 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:10Z","lastTransitionTime":"2025-11-25T13:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.225003 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675f4b11-d943-4acc-afde-6f4f45585551\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-djssv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.239106 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.253679 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.265935 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.290667 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a8fa25e9591d484f8dcb56974242924d4f31c7aef86b4376092842e57e1653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d1bfac94a5fe6b382a9a3783250d6758603d00db9ce12ea44a3b54e6018103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c91a144aef48f19006041105594111c382420f9e6556a64580f6952990ef4db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2358ac7b25088c8d1b1cc5648a1fcbfd15211826d4c47f2c98839b9924138b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9649c45bc796d60538bc1edc8f1239c7c5445de7b2f3dbf57dbe7df6ecab7d6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5dfe629efca9398d6dcc541b5315c5b6d8e036cba1c94064883785b2cc7a1a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06049a90425ba8ed3b9fff368fb341653f3ba688dee562903b75cb65faeaaf51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8700d12c2e236e704fc60136a51489d1d6890a83621480c23a6168d4a0f18f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.306404 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.320214 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.328957 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.328994 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.329005 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.329024 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.329045 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:10Z","lastTransitionTime":"2025-11-25T13:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.338591 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.431506 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.431563 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.431572 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.431588 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.431597 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:10Z","lastTransitionTime":"2025-11-25T13:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.536768 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.536809 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.536881 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.536911 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.536921 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:10Z","lastTransitionTime":"2025-11-25T13:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.551381 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-fm2cr"] Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.552060 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:10 crc kubenswrapper[4702]: E1125 13:47:10.552372 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fm2cr" podUID="88665994-5305-4de7-8b69-e38b38d5f87c" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.566808 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.579208 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.590078 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.607003 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a8fa25e9591d484f8dcb56974242924d4f31c7aef86b4376092842e57e1653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d1bfac94a5fe6b382a9a3783250d6758603d00db9ce12ea44a3b54e6018103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c91a144aef48f19006041105594111c382420f9e6556a64580f6952990ef4db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2358ac7b25088c8d1b1cc5648a1fcbfd15211826d4c47f2c98839b9924138b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9649c45bc796d60538bc1edc8f1239c7c5445de7b2f3dbf57dbe7df6ecab7d6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5dfe629efca9398d6dcc541b5315c5b6d8e036cba1c94064883785b2cc7a1a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06049a90425ba8ed3b9fff368fb341653f3ba688dee562903b75cb65faeaaf51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8700d12c2e236e704fc60136a51489d1d6890a83621480c23a6168d4a0f18f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.617217 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fm2cr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88665994-5305-4de7-8b69-e38b38d5f87c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fm2cr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.628000 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.638634 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.639035 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.639123 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.639190 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.639291 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.639352 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:10Z","lastTransitionTime":"2025-11-25T13:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.650028 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.663383 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60a7f9a29a1e574b4eb0fc72705f8c96b6f613f8a7c164605070077212854cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.675557 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.687686 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8kvv\" (UniqueName: \"kubernetes.io/projected/88665994-5305-4de7-8b69-e38b38d5f87c-kube-api-access-q8kvv\") pod \"network-metrics-daemon-fm2cr\" (UID: \"88665994-5305-4de7-8b69-e38b38d5f87c\") " pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.687786 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/88665994-5305-4de7-8b69-e38b38d5f87c-metrics-certs\") pod \"network-metrics-daemon-fm2cr\" (UID: \"88665994-5305-4de7-8b69-e38b38d5f87c\") " pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.690781 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.704192 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.715934 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.726843 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.735142 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2efcd45f14690e23b56c19e3282342933f83acd4940f89dd4046b61e4a608b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.741164 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.741211 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.741223 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.741242 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.741256 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:10Z","lastTransitionTime":"2025-11-25T13:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.744312 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675f4b11-d943-4acc-afde-6f4f45585551\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-djssv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.761362 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:10Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.789144 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8kvv\" (UniqueName: \"kubernetes.io/projected/88665994-5305-4de7-8b69-e38b38d5f87c-kube-api-access-q8kvv\") pod \"network-metrics-daemon-fm2cr\" (UID: \"88665994-5305-4de7-8b69-e38b38d5f87c\") " pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.789199 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/88665994-5305-4de7-8b69-e38b38d5f87c-metrics-certs\") pod \"network-metrics-daemon-fm2cr\" (UID: \"88665994-5305-4de7-8b69-e38b38d5f87c\") " pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:10 crc kubenswrapper[4702]: E1125 13:47:10.789320 4702 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 13:47:10 crc kubenswrapper[4702]: E1125 13:47:10.789374 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/88665994-5305-4de7-8b69-e38b38d5f87c-metrics-certs podName:88665994-5305-4de7-8b69-e38b38d5f87c nodeName:}" failed. No retries permitted until 2025-11-25 13:47:11.289356511 +0000 UTC m=+40.452006961 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/88665994-5305-4de7-8b69-e38b38d5f87c-metrics-certs") pod "network-metrics-daemon-fm2cr" (UID: "88665994-5305-4de7-8b69-e38b38d5f87c") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.800767 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.800900 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.801021 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:47:10 crc kubenswrapper[4702]: E1125 13:47:10.801061 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:47:10 crc kubenswrapper[4702]: E1125 13:47:10.801197 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:47:10 crc kubenswrapper[4702]: E1125 13:47:10.801389 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.807932 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8kvv\" (UniqueName: \"kubernetes.io/projected/88665994-5305-4de7-8b69-e38b38d5f87c-kube-api-access-q8kvv\") pod \"network-metrics-daemon-fm2cr\" (UID: \"88665994-5305-4de7-8b69-e38b38d5f87c\") " pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.843061 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.843320 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.843393 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.843458 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.843513 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:10Z","lastTransitionTime":"2025-11-25T13:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.945987 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.946027 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.946036 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.946050 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:10 crc kubenswrapper[4702]: I1125 13:47:10.946059 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:10Z","lastTransitionTime":"2025-11-25T13:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.049077 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.049163 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.049179 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.049206 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.049228 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:11Z","lastTransitionTime":"2025-11-25T13:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.087803 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" event={"ID":"675f4b11-d943-4acc-afde-6f4f45585551","Type":"ContainerStarted","Data":"62b27504ebe723ca2b16bb4fcb8c5611cffffc7a47cb356c072e2cc5aa18e796"} Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.152313 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.152351 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.152359 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.152373 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.152382 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:11Z","lastTransitionTime":"2025-11-25T13:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.255396 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.255440 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.255449 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.255463 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.255474 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:11Z","lastTransitionTime":"2025-11-25T13:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.296087 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/88665994-5305-4de7-8b69-e38b38d5f87c-metrics-certs\") pod \"network-metrics-daemon-fm2cr\" (UID: \"88665994-5305-4de7-8b69-e38b38d5f87c\") " pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:11 crc kubenswrapper[4702]: E1125 13:47:11.296261 4702 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 13:47:11 crc kubenswrapper[4702]: E1125 13:47:11.296350 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/88665994-5305-4de7-8b69-e38b38d5f87c-metrics-certs podName:88665994-5305-4de7-8b69-e38b38d5f87c nodeName:}" failed. No retries permitted until 2025-11-25 13:47:12.296332303 +0000 UTC m=+41.458982753 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/88665994-5305-4de7-8b69-e38b38d5f87c-metrics-certs") pod "network-metrics-daemon-fm2cr" (UID: "88665994-5305-4de7-8b69-e38b38d5f87c") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.358442 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.358491 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.358525 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.358541 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.358551 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:11Z","lastTransitionTime":"2025-11-25T13:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.460792 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.460829 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.460841 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.460857 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.460868 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:11Z","lastTransitionTime":"2025-11-25T13:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.563383 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.563821 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.564004 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.564223 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.564451 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:11Z","lastTransitionTime":"2025-11-25T13:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.667934 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.667981 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.667994 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.668013 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.668024 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:11Z","lastTransitionTime":"2025-11-25T13:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.771062 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.771150 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.771170 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.771200 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.771221 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:11Z","lastTransitionTime":"2025-11-25T13:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.823119 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:11Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.846146 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:11Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.864074 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:11Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.874583 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.874744 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.874807 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.874854 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.874874 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:11Z","lastTransitionTime":"2025-11-25T13:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.881643 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:11Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.897685 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60a7f9a29a1e574b4eb0fc72705f8c96b6f613f8a7c164605070077212854cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:11Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.917524 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:11Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.930860 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:11Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.950022 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:11Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.963600 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2efcd45f14690e23b56c19e3282342933f83acd4940f89dd4046b61e4a608b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:11Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.978248 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.978298 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.978339 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.978364 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.978378 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:11Z","lastTransitionTime":"2025-11-25T13:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:11 crc kubenswrapper[4702]: I1125 13:47:11.978634 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675f4b11-d943-4acc-afde-6f4f45585551\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-djssv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:11Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.005136 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:11Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.024603 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:12Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.043071 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:12Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.058562 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:12Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.071061 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:12Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.081600 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.081655 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.081686 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.081710 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.081724 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:12Z","lastTransitionTime":"2025-11-25T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.094055 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" event={"ID":"675f4b11-d943-4acc-afde-6f4f45585551","Type":"ContainerStarted","Data":"35b52152fd6304d8c2c961db7d7b144f4c17d8ee855f67e25ca711a89c60a85d"} Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.095752 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9c5v_3c8a3234-01d3-48a9-a5b4-acb46b4218dc/ovnkube-controller/0.log" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.097661 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a8fa25e9591d484f8dcb56974242924d4f31c7aef86b4376092842e57e1653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d1bfac94a5fe6b382a9a3783250d6758603d00db9ce12ea44a3b54e6018103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c91a144aef48f19006041105594111c382420f9e6556a64580f6952990ef4db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2358ac7b25088c8d1b1cc5648a1fcbfd15211826d4c47f2c98839b9924138b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9649c45bc796d60538bc1edc8f1239c7c5445de7b2f3dbf57dbe7df6ecab7d6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5dfe629efca9398d6dcc541b5315c5b6d8e036cba1c94064883785b2cc7a1a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06049a90425ba8ed3b9fff368fb341653f3ba688dee562903b75cb65faeaaf51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8700d12c2e236e704fc60136a51489d1d6890a83621480c23a6168d4a0f18f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:12Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.098255 4702 generic.go:334] "Generic (PLEG): container finished" podID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerID="06049a90425ba8ed3b9fff368fb341653f3ba688dee562903b75cb65faeaaf51" exitCode=1 Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.098331 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" event={"ID":"3c8a3234-01d3-48a9-a5b4-acb46b4218dc","Type":"ContainerDied","Data":"06049a90425ba8ed3b9fff368fb341653f3ba688dee562903b75cb65faeaaf51"} Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.099165 4702 scope.go:117] "RemoveContainer" containerID="06049a90425ba8ed3b9fff368fb341653f3ba688dee562903b75cb65faeaaf51" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.113274 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fm2cr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88665994-5305-4de7-8b69-e38b38d5f87c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fm2cr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:12Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.128241 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:12Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.142547 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:12Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.156433 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2efcd45f14690e23b56c19e3282342933f83acd4940f89dd4046b61e4a608b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:12Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.173546 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675f4b11-d943-4acc-afde-6f4f45585551\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-djssv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:12Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.186484 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.186824 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.186917 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.187036 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.187116 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:12Z","lastTransitionTime":"2025-11-25T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.202155 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:12Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.224324 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:12Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.246004 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:12Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.269620 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:12Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.287047 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:12Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.290267 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.290309 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.290322 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.290363 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.290376 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:12Z","lastTransitionTime":"2025-11-25T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.306524 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/88665994-5305-4de7-8b69-e38b38d5f87c-metrics-certs\") pod \"network-metrics-daemon-fm2cr\" (UID: \"88665994-5305-4de7-8b69-e38b38d5f87c\") " pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:12 crc kubenswrapper[4702]: E1125 13:47:12.306898 4702 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 13:47:12 crc kubenswrapper[4702]: E1125 13:47:12.307100 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/88665994-5305-4de7-8b69-e38b38d5f87c-metrics-certs podName:88665994-5305-4de7-8b69-e38b38d5f87c nodeName:}" failed. No retries permitted until 2025-11-25 13:47:14.307048858 +0000 UTC m=+43.469699468 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/88665994-5305-4de7-8b69-e38b38d5f87c-metrics-certs") pod "network-metrics-daemon-fm2cr" (UID: "88665994-5305-4de7-8b69-e38b38d5f87c") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.320645 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a8fa25e9591d484f8dcb56974242924d4f31c7aef86b4376092842e57e1653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d1bfac94a5fe6b382a9a3783250d6758603d00db9ce12ea44a3b54e6018103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c91a144aef48f19006041105594111c382420f9e6556a64580f6952990ef4db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2358ac7b25088c8d1b1cc5648a1fcbfd15211826d4c47f2c98839b9924138b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9649c45bc796d60538bc1edc8f1239c7c5445de7b2f3dbf57dbe7df6ecab7d6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5dfe629efca9398d6dcc541b5315c5b6d8e036cba1c94064883785b2cc7a1a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06049a90425ba8ed3b9fff368fb341653f3ba688dee562903b75cb65faeaaf51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06049a90425ba8ed3b9fff368fb341653f3ba688dee562903b75cb65faeaaf51\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T13:47:11Z\\\",\\\"message\\\":\\\" 5950 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 13:47:10.533803 5950 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 13:47:10.533809 5950 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 13:47:10.533814 5950 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 13:47:10.533821 5950 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 13:47:10.534115 5950 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 13:47:10.534135 5950 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 13:47:10.534314 5950 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 13:47:10.534344 5950 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 13:47:10.534324 5950 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 13:47:10.534406 5950 factory.go:656] Stopping watch factory\\\\nI1125 13:47:10.534412 5950 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 13:47:10.534458 5950 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 13:47:10.534491 5950 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 13:47:10.534772 5950 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8700d12c2e236e704fc60136a51489d1d6890a83621480c23a6168d4a0f18f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:12Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.339038 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fm2cr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88665994-5305-4de7-8b69-e38b38d5f87c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fm2cr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:12Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.357020 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:12Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.381025 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:12Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.394068 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.394157 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.394184 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.394299 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.394325 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:12Z","lastTransitionTime":"2025-11-25T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.403813 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:12Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.420936 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:12Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.438790 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:12Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.460149 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60a7f9a29a1e574b4eb0fc72705f8c96b6f613f8a7c164605070077212854cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:12Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.498336 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.498394 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.498415 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.498447 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.498468 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:12Z","lastTransitionTime":"2025-11-25T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.603147 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.603552 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.603745 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.604862 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.605018 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:12Z","lastTransitionTime":"2025-11-25T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.709043 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.709342 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.709428 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.709535 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.709597 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:12Z","lastTransitionTime":"2025-11-25T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.799982 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.800096 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.800144 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.800203 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:47:12 crc kubenswrapper[4702]: E1125 13:47:12.800303 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:47:12 crc kubenswrapper[4702]: E1125 13:47:12.800541 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:47:12 crc kubenswrapper[4702]: E1125 13:47:12.800720 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:47:12 crc kubenswrapper[4702]: E1125 13:47:12.800887 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fm2cr" podUID="88665994-5305-4de7-8b69-e38b38d5f87c" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.813458 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.813542 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.813573 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.813605 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.813625 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:12Z","lastTransitionTime":"2025-11-25T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.916268 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.916392 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.916405 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.916420 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:12 crc kubenswrapper[4702]: I1125 13:47:12.916430 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:12Z","lastTransitionTime":"2025-11-25T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.019700 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.019779 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.019798 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.019829 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.019851 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:13Z","lastTransitionTime":"2025-11-25T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.123290 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.123819 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.123834 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.123860 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.123905 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:13Z","lastTransitionTime":"2025-11-25T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.226884 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.226920 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.226928 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.226944 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.226952 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:13Z","lastTransitionTime":"2025-11-25T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.330913 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.330984 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.331018 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.331049 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.331071 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:13Z","lastTransitionTime":"2025-11-25T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.436599 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.436655 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.436668 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.436694 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.436707 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:13Z","lastTransitionTime":"2025-11-25T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.538871 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.538914 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.538927 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.538944 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.538953 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:13Z","lastTransitionTime":"2025-11-25T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.641160 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.641202 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.641212 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.641226 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.641252 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:13Z","lastTransitionTime":"2025-11-25T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.747515 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.747563 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.747578 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.747599 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.747609 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:13Z","lastTransitionTime":"2025-11-25T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.849508 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.849549 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.849557 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.849570 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.849579 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:13Z","lastTransitionTime":"2025-11-25T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.951935 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.951975 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.951984 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.952001 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:13 crc kubenswrapper[4702]: I1125 13:47:13.952010 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:13Z","lastTransitionTime":"2025-11-25T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.054570 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.055093 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.055333 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.055548 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.055719 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:14Z","lastTransitionTime":"2025-11-25T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.108002 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9c5v_3c8a3234-01d3-48a9-a5b4-acb46b4218dc/ovnkube-controller/0.log" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.111803 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" event={"ID":"3c8a3234-01d3-48a9-a5b4-acb46b4218dc","Type":"ContainerStarted","Data":"427ca923a5d36fe0441dbe5e9dce7e39c374164534025c436e743e2c64a63b1c"} Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.112196 4702 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.113971 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" event={"ID":"675f4b11-d943-4acc-afde-6f4f45585551","Type":"ContainerStarted","Data":"f010c83c609ad0554ed60899b53504cd7d859528a4947f45356466d15930c8ee"} Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.132811 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.146682 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.158713 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.158752 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.158764 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.158783 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.158795 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:14Z","lastTransitionTime":"2025-11-25T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.166126 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.183865 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.205250 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2efcd45f14690e23b56c19e3282342933f83acd4940f89dd4046b61e4a608b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.222100 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675f4b11-d943-4acc-afde-6f4f45585551\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-djssv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.243679 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.256730 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.260823 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.260866 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.260877 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.260899 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.260910 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:14Z","lastTransitionTime":"2025-11-25T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.271488 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.285387 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.309400 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a8fa25e9591d484f8dcb56974242924d4f31c7aef86b4376092842e57e1653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d1bfac94a5fe6b382a9a3783250d6758603d00db9ce12ea44a3b54e6018103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c91a144aef48f19006041105594111c382420f9e6556a64580f6952990ef4db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2358ac7b25088c8d1b1cc5648a1fcbfd15211826d4c47f2c98839b9924138b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9649c45bc796d60538bc1edc8f1239c7c5445de7b2f3dbf57dbe7df6ecab7d6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5dfe629efca9398d6dcc541b5315c5b6d8e036cba1c94064883785b2cc7a1a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://427ca923a5d36fe0441dbe5e9dce7e39c374164534025c436e743e2c64a63b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06049a90425ba8ed3b9fff368fb341653f3ba688dee562903b75cb65faeaaf51\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T13:47:11Z\\\",\\\"message\\\":\\\" 5950 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 13:47:10.533803 5950 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 13:47:10.533809 5950 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 13:47:10.533814 5950 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 13:47:10.533821 5950 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 13:47:10.534115 5950 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 13:47:10.534135 5950 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 13:47:10.534314 5950 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 13:47:10.534344 5950 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 13:47:10.534324 5950 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 13:47:10.534406 5950 factory.go:656] Stopping watch factory\\\\nI1125 13:47:10.534412 5950 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 13:47:10.534458 5950 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 13:47:10.534491 5950 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 13:47:10.534772 5950 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8700d12c2e236e704fc60136a51489d1d6890a83621480c23a6168d4a0f18f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.326143 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fm2cr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88665994-5305-4de7-8b69-e38b38d5f87c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fm2cr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.344526 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.348594 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/88665994-5305-4de7-8b69-e38b38d5f87c-metrics-certs\") pod \"network-metrics-daemon-fm2cr\" (UID: \"88665994-5305-4de7-8b69-e38b38d5f87c\") " pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:14 crc kubenswrapper[4702]: E1125 13:47:14.348704 4702 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 13:47:14 crc kubenswrapper[4702]: E1125 13:47:14.348752 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/88665994-5305-4de7-8b69-e38b38d5f87c-metrics-certs podName:88665994-5305-4de7-8b69-e38b38d5f87c nodeName:}" failed. No retries permitted until 2025-11-25 13:47:18.348738522 +0000 UTC m=+47.511388972 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/88665994-5305-4de7-8b69-e38b38d5f87c-metrics-certs") pod "network-metrics-daemon-fm2cr" (UID: "88665994-5305-4de7-8b69-e38b38d5f87c") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.357901 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.363352 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.363404 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.363419 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.363440 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.363452 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:14Z","lastTransitionTime":"2025-11-25T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.372362 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.388296 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60a7f9a29a1e574b4eb0fc72705f8c96b6f613f8a7c164605070077212854cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.401560 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.416098 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.428490 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.445001 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.456360 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2efcd45f14690e23b56c19e3282342933f83acd4940f89dd4046b61e4a608b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.466984 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.467025 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.467034 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.467054 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.467067 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:14Z","lastTransitionTime":"2025-11-25T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.471663 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675f4b11-d943-4acc-afde-6f4f45585551\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35b52152fd6304d8c2c961db7d7b144f4c17d8ee855f67e25ca711a89c60a85d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010c83c609ad0554ed60899b53504cd7d859528a4947f45356466d15930c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-djssv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.539645 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.554920 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.567384 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.568993 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.569028 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.569040 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.569058 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.569070 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:14Z","lastTransitionTime":"2025-11-25T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.583876 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.597093 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.616759 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a8fa25e9591d484f8dcb56974242924d4f31c7aef86b4376092842e57e1653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d1bfac94a5fe6b382a9a3783250d6758603d00db9ce12ea44a3b54e6018103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c91a144aef48f19006041105594111c382420f9e6556a64580f6952990ef4db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2358ac7b25088c8d1b1cc5648a1fcbfd15211826d4c47f2c98839b9924138b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9649c45bc796d60538bc1edc8f1239c7c5445de7b2f3dbf57dbe7df6ecab7d6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5dfe629efca9398d6dcc541b5315c5b6d8e036cba1c94064883785b2cc7a1a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://427ca923a5d36fe0441dbe5e9dce7e39c374164534025c436e743e2c64a63b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06049a90425ba8ed3b9fff368fb341653f3ba688dee562903b75cb65faeaaf51\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T13:47:11Z\\\",\\\"message\\\":\\\" 5950 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 13:47:10.533803 5950 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 13:47:10.533809 5950 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 13:47:10.533814 5950 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 13:47:10.533821 5950 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 13:47:10.534115 5950 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 13:47:10.534135 5950 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 13:47:10.534314 5950 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 13:47:10.534344 5950 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 13:47:10.534324 5950 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 13:47:10.534406 5950 factory.go:656] Stopping watch factory\\\\nI1125 13:47:10.534412 5950 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 13:47:10.534458 5950 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 13:47:10.534491 5950 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 13:47:10.534772 5950 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8700d12c2e236e704fc60136a51489d1d6890a83621480c23a6168d4a0f18f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.628907 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fm2cr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88665994-5305-4de7-8b69-e38b38d5f87c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fm2cr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.640039 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.651128 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.664176 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.671178 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.671232 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.671259 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.671279 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.671293 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:14Z","lastTransitionTime":"2025-11-25T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.678037 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.698083 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60a7f9a29a1e574b4eb0fc72705f8c96b6f613f8a7c164605070077212854cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.773866 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.773906 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.773918 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.773935 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.773945 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:14Z","lastTransitionTime":"2025-11-25T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.800755 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.800790 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.800826 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.800755 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:14 crc kubenswrapper[4702]: E1125 13:47:14.800951 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:47:14 crc kubenswrapper[4702]: E1125 13:47:14.800869 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:47:14 crc kubenswrapper[4702]: E1125 13:47:14.801100 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:47:14 crc kubenswrapper[4702]: E1125 13:47:14.801184 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fm2cr" podUID="88665994-5305-4de7-8b69-e38b38d5f87c" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.877069 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.877108 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.877118 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.877133 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.877144 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:14Z","lastTransitionTime":"2025-11-25T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.980056 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.980091 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.980100 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.980112 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:14 crc kubenswrapper[4702]: I1125 13:47:14.980121 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:14Z","lastTransitionTime":"2025-11-25T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.083477 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.083539 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.083553 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.083580 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.083597 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:15Z","lastTransitionTime":"2025-11-25T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.123515 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9c5v_3c8a3234-01d3-48a9-a5b4-acb46b4218dc/ovnkube-controller/1.log" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.124820 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9c5v_3c8a3234-01d3-48a9-a5b4-acb46b4218dc/ovnkube-controller/0.log" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.128956 4702 generic.go:334] "Generic (PLEG): container finished" podID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerID="427ca923a5d36fe0441dbe5e9dce7e39c374164534025c436e743e2c64a63b1c" exitCode=1 Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.129308 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" event={"ID":"3c8a3234-01d3-48a9-a5b4-acb46b4218dc","Type":"ContainerDied","Data":"427ca923a5d36fe0441dbe5e9dce7e39c374164534025c436e743e2c64a63b1c"} Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.129476 4702 scope.go:117] "RemoveContainer" containerID="06049a90425ba8ed3b9fff368fb341653f3ba688dee562903b75cb65faeaaf51" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.130848 4702 scope.go:117] "RemoveContainer" containerID="427ca923a5d36fe0441dbe5e9dce7e39c374164534025c436e743e2c64a63b1c" Nov 25 13:47:15 crc kubenswrapper[4702]: E1125 13:47:15.131192 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-p9c5v_openshift-ovn-kubernetes(3c8a3234-01d3-48a9-a5b4-acb46b4218dc)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.149804 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.165516 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.184451 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.185691 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.185813 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.185896 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.185941 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.185952 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:15Z","lastTransitionTime":"2025-11-25T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.197485 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2efcd45f14690e23b56c19e3282342933f83acd4940f89dd4046b61e4a608b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.211882 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675f4b11-d943-4acc-afde-6f4f45585551\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35b52152fd6304d8c2c961db7d7b144f4c17d8ee855f67e25ca711a89c60a85d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010c83c609ad0554ed60899b53504cd7d859528a4947f45356466d15930c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-djssv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.233997 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.255363 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.272030 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.285823 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.288435 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.288502 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.288542 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.288571 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.288593 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:15Z","lastTransitionTime":"2025-11-25T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.300477 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.323294 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a8fa25e9591d484f8dcb56974242924d4f31c7aef86b4376092842e57e1653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d1bfac94a5fe6b382a9a3783250d6758603d00db9ce12ea44a3b54e6018103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c91a144aef48f19006041105594111c382420f9e6556a64580f6952990ef4db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2358ac7b25088c8d1b1cc5648a1fcbfd15211826d4c47f2c98839b9924138b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9649c45bc796d60538bc1edc8f1239c7c5445de7b2f3dbf57dbe7df6ecab7d6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5dfe629efca9398d6dcc541b5315c5b6d8e036cba1c94064883785b2cc7a1a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://427ca923a5d36fe0441dbe5e9dce7e39c374164534025c436e743e2c64a63b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06049a90425ba8ed3b9fff368fb341653f3ba688dee562903b75cb65faeaaf51\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T13:47:11Z\\\",\\\"message\\\":\\\" 5950 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 13:47:10.533803 5950 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 13:47:10.533809 5950 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 13:47:10.533814 5950 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 13:47:10.533821 5950 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 13:47:10.534115 5950 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 13:47:10.534135 5950 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 13:47:10.534314 5950 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 13:47:10.534344 5950 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 13:47:10.534324 5950 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 13:47:10.534406 5950 factory.go:656] Stopping watch factory\\\\nI1125 13:47:10.534412 5950 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 13:47:10.534458 5950 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 13:47:10.534491 5950 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 13:47:10.534772 5950 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://427ca923a5d36fe0441dbe5e9dce7e39c374164534025c436e743e2c64a63b1c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"message\\\":\\\":134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv in node crc\\\\nF1125 13:47:14.584282 6213 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z]\\\\nI1125 13:47:14.584350 6213 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv after 0 failed attempt(s)\\\\nI1125 13:47:14.584351 6213 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-qqr76 in node crc\\\\nI1125 13:47:14.584357 6213 default_network_controller.go:776] Recording success event on pod\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8700d12c2e236e704fc60136a51489d1d6890a83621480c23a6168d4a0f18f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.335497 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fm2cr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88665994-5305-4de7-8b69-e38b38d5f87c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fm2cr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.349854 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.366949 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.391526 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.391594 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.391613 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.391640 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.391661 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:15Z","lastTransitionTime":"2025-11-25T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.394270 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.408739 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.422594 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60a7f9a29a1e574b4eb0fc72705f8c96b6f613f8a7c164605070077212854cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.494514 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.494802 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.494925 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.495334 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.495458 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:15Z","lastTransitionTime":"2025-11-25T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.598420 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.598458 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.598469 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.598489 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.598500 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:15Z","lastTransitionTime":"2025-11-25T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.701177 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.701210 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.701219 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.701254 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.701264 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:15Z","lastTransitionTime":"2025-11-25T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.803989 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.804035 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.804043 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.804057 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.804066 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:15Z","lastTransitionTime":"2025-11-25T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.906666 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.906714 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.906725 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.906739 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:15 crc kubenswrapper[4702]: I1125 13:47:15.906751 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:15Z","lastTransitionTime":"2025-11-25T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.009511 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.009547 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.009555 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.009571 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.009581 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:16Z","lastTransitionTime":"2025-11-25T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.027372 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.112931 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.112974 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.112988 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.113019 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.113031 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:16Z","lastTransitionTime":"2025-11-25T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.133720 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9c5v_3c8a3234-01d3-48a9-a5b4-acb46b4218dc/ovnkube-controller/1.log" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.136700 4702 scope.go:117] "RemoveContainer" containerID="427ca923a5d36fe0441dbe5e9dce7e39c374164534025c436e743e2c64a63b1c" Nov 25 13:47:16 crc kubenswrapper[4702]: E1125 13:47:16.136863 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-p9c5v_openshift-ovn-kubernetes(3c8a3234-01d3-48a9-a5b4-acb46b4218dc)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.162924 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:16Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.176702 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:16Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.188771 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:16Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.200539 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:16Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.212567 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:16Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.215794 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.215844 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.215854 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.215875 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.215886 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:16Z","lastTransitionTime":"2025-11-25T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.225707 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2efcd45f14690e23b56c19e3282342933f83acd4940f89dd4046b61e4a608b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:16Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.243419 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675f4b11-d943-4acc-afde-6f4f45585551\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35b52152fd6304d8c2c961db7d7b144f4c17d8ee855f67e25ca711a89c60a85d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010c83c609ad0554ed60899b53504cd7d859528a4947f45356466d15930c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-djssv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:16Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.259676 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:16Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.273799 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:16Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.286382 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:16Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.303578 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a8fa25e9591d484f8dcb56974242924d4f31c7aef86b4376092842e57e1653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d1bfac94a5fe6b382a9a3783250d6758603d00db9ce12ea44a3b54e6018103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c91a144aef48f19006041105594111c382420f9e6556a64580f6952990ef4db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2358ac7b25088c8d1b1cc5648a1fcbfd15211826d4c47f2c98839b9924138b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9649c45bc796d60538bc1edc8f1239c7c5445de7b2f3dbf57dbe7df6ecab7d6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5dfe629efca9398d6dcc541b5315c5b6d8e036cba1c94064883785b2cc7a1a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://427ca923a5d36fe0441dbe5e9dce7e39c374164534025c436e743e2c64a63b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://427ca923a5d36fe0441dbe5e9dce7e39c374164534025c436e743e2c64a63b1c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"message\\\":\\\":134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv in node crc\\\\nF1125 13:47:14.584282 6213 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z]\\\\nI1125 13:47:14.584350 6213 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv after 0 failed attempt(s)\\\\nI1125 13:47:14.584351 6213 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-qqr76 in node crc\\\\nI1125 13:47:14.584357 6213 default_network_controller.go:776] Recording success event on pod\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-p9c5v_openshift-ovn-kubernetes(3c8a3234-01d3-48a9-a5b4-acb46b4218dc)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8700d12c2e236e704fc60136a51489d1d6890a83621480c23a6168d4a0f18f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:16Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.314626 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fm2cr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88665994-5305-4de7-8b69-e38b38d5f87c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fm2cr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:16Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.319079 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.319129 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.319143 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.319168 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.319182 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:16Z","lastTransitionTime":"2025-11-25T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.331438 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:16Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.345814 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:16Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.359664 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:16Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.374439 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:16Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.391155 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60a7f9a29a1e574b4eb0fc72705f8c96b6f613f8a7c164605070077212854cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:16Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.422093 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.422144 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.422156 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.422177 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.422189 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:16Z","lastTransitionTime":"2025-11-25T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.525207 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.525338 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.525365 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.525404 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.525422 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:16Z","lastTransitionTime":"2025-11-25T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.628067 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.628115 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.628128 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.628144 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.628158 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:16Z","lastTransitionTime":"2025-11-25T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.730634 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.730673 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.730682 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.730697 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.730710 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:16Z","lastTransitionTime":"2025-11-25T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.800820 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:47:16 crc kubenswrapper[4702]: E1125 13:47:16.800973 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.801043 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.801062 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:16 crc kubenswrapper[4702]: E1125 13:47:16.801141 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fm2cr" podUID="88665994-5305-4de7-8b69-e38b38d5f87c" Nov 25 13:47:16 crc kubenswrapper[4702]: E1125 13:47:16.801306 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.801813 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:47:16 crc kubenswrapper[4702]: E1125 13:47:16.802059 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.833291 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.833398 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.833420 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.833454 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.833476 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:16Z","lastTransitionTime":"2025-11-25T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.937226 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.937295 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.937304 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.937324 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:16 crc kubenswrapper[4702]: I1125 13:47:16.937335 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:16Z","lastTransitionTime":"2025-11-25T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.039693 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.039740 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.039758 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.039774 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.039786 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:17Z","lastTransitionTime":"2025-11-25T13:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.143412 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.143808 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.143942 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.144046 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.144144 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:17Z","lastTransitionTime":"2025-11-25T13:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.247661 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.247727 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.247747 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.247776 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.247797 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:17Z","lastTransitionTime":"2025-11-25T13:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.337567 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.337636 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.337670 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.337696 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.337714 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:17Z","lastTransitionTime":"2025-11-25T13:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:17 crc kubenswrapper[4702]: E1125 13:47:17.355163 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"96cc93a3-66f3-477f-bbc8-ef3a984ed406\\\",\\\"systemUUID\\\":\\\"bb366f18-cc3c-451f-9d6a-c0cc41864b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:17Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.360098 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.360319 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.360412 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.360549 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.360627 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:17Z","lastTransitionTime":"2025-11-25T13:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:17 crc kubenswrapper[4702]: E1125 13:47:17.376125 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"96cc93a3-66f3-477f-bbc8-ef3a984ed406\\\",\\\"systemUUID\\\":\\\"bb366f18-cc3c-451f-9d6a-c0cc41864b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:17Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.382011 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.382082 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.382104 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.382134 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.382154 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:17Z","lastTransitionTime":"2025-11-25T13:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:17 crc kubenswrapper[4702]: E1125 13:47:17.397889 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"96cc93a3-66f3-477f-bbc8-ef3a984ed406\\\",\\\"systemUUID\\\":\\\"bb366f18-cc3c-451f-9d6a-c0cc41864b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:17Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.403190 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.403614 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.403701 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.403881 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.403960 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:17Z","lastTransitionTime":"2025-11-25T13:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:17 crc kubenswrapper[4702]: E1125 13:47:17.421227 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"96cc93a3-66f3-477f-bbc8-ef3a984ed406\\\",\\\"systemUUID\\\":\\\"bb366f18-cc3c-451f-9d6a-c0cc41864b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:17Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.425429 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.425492 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.425509 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.425532 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.425546 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:17Z","lastTransitionTime":"2025-11-25T13:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:17 crc kubenswrapper[4702]: E1125 13:47:17.440561 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"96cc93a3-66f3-477f-bbc8-ef3a984ed406\\\",\\\"systemUUID\\\":\\\"bb366f18-cc3c-451f-9d6a-c0cc41864b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:17Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:17 crc kubenswrapper[4702]: E1125 13:47:17.440747 4702 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.442926 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.442968 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.442982 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.443006 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.443028 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:17Z","lastTransitionTime":"2025-11-25T13:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.546383 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.546435 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.546445 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.546483 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.546498 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:17Z","lastTransitionTime":"2025-11-25T13:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.649396 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.649450 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.649463 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.649483 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.649496 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:17Z","lastTransitionTime":"2025-11-25T13:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.753540 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.753627 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.753648 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.753680 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.753707 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:17Z","lastTransitionTime":"2025-11-25T13:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.857835 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.857882 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.857901 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.857926 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.857949 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:17Z","lastTransitionTime":"2025-11-25T13:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.960620 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.960704 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.960713 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.960732 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:17 crc kubenswrapper[4702]: I1125 13:47:17.960742 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:17Z","lastTransitionTime":"2025-11-25T13:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.064841 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.064892 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.064905 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.064926 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.064944 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:18Z","lastTransitionTime":"2025-11-25T13:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.167738 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.167806 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.167822 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.167846 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.167863 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:18Z","lastTransitionTime":"2025-11-25T13:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.271207 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.271293 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.271308 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.271334 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.271348 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:18Z","lastTransitionTime":"2025-11-25T13:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.374032 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.374086 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.374098 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.374126 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.374140 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:18Z","lastTransitionTime":"2025-11-25T13:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.412511 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/88665994-5305-4de7-8b69-e38b38d5f87c-metrics-certs\") pod \"network-metrics-daemon-fm2cr\" (UID: \"88665994-5305-4de7-8b69-e38b38d5f87c\") " pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:18 crc kubenswrapper[4702]: E1125 13:47:18.412716 4702 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 13:47:18 crc kubenswrapper[4702]: E1125 13:47:18.412991 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/88665994-5305-4de7-8b69-e38b38d5f87c-metrics-certs podName:88665994-5305-4de7-8b69-e38b38d5f87c nodeName:}" failed. No retries permitted until 2025-11-25 13:47:26.412964112 +0000 UTC m=+55.575614572 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/88665994-5305-4de7-8b69-e38b38d5f87c-metrics-certs") pod "network-metrics-daemon-fm2cr" (UID: "88665994-5305-4de7-8b69-e38b38d5f87c") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.477730 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.477782 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.477797 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.477816 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.477827 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:18Z","lastTransitionTime":"2025-11-25T13:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.581208 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.581285 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.581296 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.581318 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.581332 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:18Z","lastTransitionTime":"2025-11-25T13:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.684378 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.684439 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.684452 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.684466 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.684499 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:18Z","lastTransitionTime":"2025-11-25T13:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.787223 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.787284 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.787293 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.787307 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.787318 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:18Z","lastTransitionTime":"2025-11-25T13:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.801143 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.801272 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.801188 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.801166 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:18 crc kubenswrapper[4702]: E1125 13:47:18.801442 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:47:18 crc kubenswrapper[4702]: E1125 13:47:18.801565 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fm2cr" podUID="88665994-5305-4de7-8b69-e38b38d5f87c" Nov 25 13:47:18 crc kubenswrapper[4702]: E1125 13:47:18.801706 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:47:18 crc kubenswrapper[4702]: E1125 13:47:18.801791 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.890318 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.890377 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.890390 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.890411 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.890428 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:18Z","lastTransitionTime":"2025-11-25T13:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.994282 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.994337 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.994348 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.994371 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:18 crc kubenswrapper[4702]: I1125 13:47:18.994421 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:18Z","lastTransitionTime":"2025-11-25T13:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.098288 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.098345 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.098361 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.098380 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.098394 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:19Z","lastTransitionTime":"2025-11-25T13:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.202050 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.202098 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.202107 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.202123 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.202133 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:19Z","lastTransitionTime":"2025-11-25T13:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.305801 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.305868 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.305886 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.305914 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.305934 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:19Z","lastTransitionTime":"2025-11-25T13:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.408913 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.408984 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.408999 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.409025 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.409046 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:19Z","lastTransitionTime":"2025-11-25T13:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.512959 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.513020 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.513042 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.513063 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.513078 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:19Z","lastTransitionTime":"2025-11-25T13:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.615681 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.615764 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.615783 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.615812 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.615831 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:19Z","lastTransitionTime":"2025-11-25T13:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.719183 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.719300 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.719320 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.719349 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.719382 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:19Z","lastTransitionTime":"2025-11-25T13:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.822485 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.822566 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.822578 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.822600 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.822614 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:19Z","lastTransitionTime":"2025-11-25T13:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.926189 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.926269 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.926283 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.926311 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:19 crc kubenswrapper[4702]: I1125 13:47:19.926343 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:19Z","lastTransitionTime":"2025-11-25T13:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.029444 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.029504 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.029513 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.029532 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.029549 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:20Z","lastTransitionTime":"2025-11-25T13:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.133038 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.133086 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.133099 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.133118 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.133132 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:20Z","lastTransitionTime":"2025-11-25T13:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.236548 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.236614 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.236632 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.236655 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.236672 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:20Z","lastTransitionTime":"2025-11-25T13:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.339732 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.339784 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.339793 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.339808 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.339844 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:20Z","lastTransitionTime":"2025-11-25T13:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.442560 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.442594 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.442602 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.442616 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.442626 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:20Z","lastTransitionTime":"2025-11-25T13:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.545782 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.545832 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.545844 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.545862 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.545875 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:20Z","lastTransitionTime":"2025-11-25T13:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.647900 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.647940 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.647951 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.647991 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.648023 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:20Z","lastTransitionTime":"2025-11-25T13:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.750366 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.750420 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.750431 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.750453 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.750466 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:20Z","lastTransitionTime":"2025-11-25T13:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.801202 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.801215 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.801297 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:47:20 crc kubenswrapper[4702]: E1125 13:47:20.801368 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.801440 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:20 crc kubenswrapper[4702]: E1125 13:47:20.801522 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:47:20 crc kubenswrapper[4702]: E1125 13:47:20.801630 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fm2cr" podUID="88665994-5305-4de7-8b69-e38b38d5f87c" Nov 25 13:47:20 crc kubenswrapper[4702]: E1125 13:47:20.801792 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.853909 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.854002 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.854022 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.854050 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.854070 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:20Z","lastTransitionTime":"2025-11-25T13:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.956622 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.956721 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.956738 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.956767 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:20 crc kubenswrapper[4702]: I1125 13:47:20.956783 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:20Z","lastTransitionTime":"2025-11-25T13:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.059209 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.059567 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.059585 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.059603 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.059614 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:21Z","lastTransitionTime":"2025-11-25T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.161860 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.161910 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.161921 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.161937 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.161947 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:21Z","lastTransitionTime":"2025-11-25T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.264664 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.264705 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.264715 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.264733 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.264748 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:21Z","lastTransitionTime":"2025-11-25T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.367824 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.367882 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.367893 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.367906 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.367917 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:21Z","lastTransitionTime":"2025-11-25T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.470580 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.470636 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.470650 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.470673 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.470689 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:21Z","lastTransitionTime":"2025-11-25T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.573491 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.573548 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.573564 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.573619 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.573636 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:21Z","lastTransitionTime":"2025-11-25T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.676604 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.676692 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.676711 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.676739 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.676757 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:21Z","lastTransitionTime":"2025-11-25T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.779151 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.779204 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.779219 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.779260 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.779271 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:21Z","lastTransitionTime":"2025-11-25T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.820806 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.834469 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2efcd45f14690e23b56c19e3282342933f83acd4940f89dd4046b61e4a608b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.848397 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675f4b11-d943-4acc-afde-6f4f45585551\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35b52152fd6304d8c2c961db7d7b144f4c17d8ee855f67e25ca711a89c60a85d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010c83c609ad0554ed60899b53504cd7d859528a4947f45356466d15930c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-djssv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.870137 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.881840 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.881885 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.881899 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.881922 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.881938 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:21Z","lastTransitionTime":"2025-11-25T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.886595 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.904303 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.917170 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.932123 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.954845 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a8fa25e9591d484f8dcb56974242924d4f31c7aef86b4376092842e57e1653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d1bfac94a5fe6b382a9a3783250d6758603d00db9ce12ea44a3b54e6018103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c91a144aef48f19006041105594111c382420f9e6556a64580f6952990ef4db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2358ac7b25088c8d1b1cc5648a1fcbfd15211826d4c47f2c98839b9924138b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9649c45bc796d60538bc1edc8f1239c7c5445de7b2f3dbf57dbe7df6ecab7d6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5dfe629efca9398d6dcc541b5315c5b6d8e036cba1c94064883785b2cc7a1a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://427ca923a5d36fe0441dbe5e9dce7e39c374164534025c436e743e2c64a63b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://427ca923a5d36fe0441dbe5e9dce7e39c374164534025c436e743e2c64a63b1c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"message\\\":\\\":134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv in node crc\\\\nF1125 13:47:14.584282 6213 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z]\\\\nI1125 13:47:14.584350 6213 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv after 0 failed attempt(s)\\\\nI1125 13:47:14.584351 6213 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-qqr76 in node crc\\\\nI1125 13:47:14.584357 6213 default_network_controller.go:776] Recording success event on pod\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-p9c5v_openshift-ovn-kubernetes(3c8a3234-01d3-48a9-a5b4-acb46b4218dc)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8700d12c2e236e704fc60136a51489d1d6890a83621480c23a6168d4a0f18f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.970202 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fm2cr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88665994-5305-4de7-8b69-e38b38d5f87c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fm2cr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.984047 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.984100 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.984126 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.984141 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.984151 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:21Z","lastTransitionTime":"2025-11-25T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:21 crc kubenswrapper[4702]: I1125 13:47:21.987094 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.001089 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:21Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.016631 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:22Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.032585 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:22Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.046847 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:22Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.062574 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:22Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.080714 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60a7f9a29a1e574b4eb0fc72705f8c96b6f613f8a7c164605070077212854cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:22Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.086869 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.086906 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.086917 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.086931 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.086942 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:22Z","lastTransitionTime":"2025-11-25T13:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.190493 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.190541 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.190550 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.190567 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.190576 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:22Z","lastTransitionTime":"2025-11-25T13:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.293602 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.293663 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.293676 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.293699 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.293715 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:22Z","lastTransitionTime":"2025-11-25T13:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.396640 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.396719 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.396729 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.396741 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.396749 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:22Z","lastTransitionTime":"2025-11-25T13:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.498913 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.498948 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.498957 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.498970 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.498979 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:22Z","lastTransitionTime":"2025-11-25T13:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.556512 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.556587 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.556615 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:22 crc kubenswrapper[4702]: E1125 13:47:22.556646 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:47:54.556629239 +0000 UTC m=+83.719279689 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:47:22 crc kubenswrapper[4702]: E1125 13:47:22.556682 4702 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 13:47:22 crc kubenswrapper[4702]: E1125 13:47:22.556720 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 13:47:54.556711491 +0000 UTC m=+83.719361941 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 13:47:22 crc kubenswrapper[4702]: E1125 13:47:22.556809 4702 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 13:47:22 crc kubenswrapper[4702]: E1125 13:47:22.556896 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 13:47:54.556879465 +0000 UTC m=+83.719529915 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.601794 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.601840 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.601850 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.601870 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.601882 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:22Z","lastTransitionTime":"2025-11-25T13:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.657686 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.657752 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:47:22 crc kubenswrapper[4702]: E1125 13:47:22.657895 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 13:47:22 crc kubenswrapper[4702]: E1125 13:47:22.657911 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 13:47:22 crc kubenswrapper[4702]: E1125 13:47:22.657921 4702 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 13:47:22 crc kubenswrapper[4702]: E1125 13:47:22.657931 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 13:47:22 crc kubenswrapper[4702]: E1125 13:47:22.657973 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 13:47:54.657959601 +0000 UTC m=+83.820610051 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 13:47:22 crc kubenswrapper[4702]: E1125 13:47:22.657981 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 13:47:22 crc kubenswrapper[4702]: E1125 13:47:22.657996 4702 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 13:47:22 crc kubenswrapper[4702]: E1125 13:47:22.658050 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 13:47:54.658031713 +0000 UTC m=+83.820682233 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.704053 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.704083 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.704091 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.704104 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.704113 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:22Z","lastTransitionTime":"2025-11-25T13:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.800548 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.800605 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.800561 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.800562 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:22 crc kubenswrapper[4702]: E1125 13:47:22.800690 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:47:22 crc kubenswrapper[4702]: E1125 13:47:22.800803 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fm2cr" podUID="88665994-5305-4de7-8b69-e38b38d5f87c" Nov 25 13:47:22 crc kubenswrapper[4702]: E1125 13:47:22.800902 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:47:22 crc kubenswrapper[4702]: E1125 13:47:22.800928 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.806254 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.806284 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.806292 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.806305 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.806315 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:22Z","lastTransitionTime":"2025-11-25T13:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.909052 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.909096 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.909103 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.909115 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:22 crc kubenswrapper[4702]: I1125 13:47:22.909125 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:22Z","lastTransitionTime":"2025-11-25T13:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.011687 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.011749 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.011767 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.011789 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.011806 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:23Z","lastTransitionTime":"2025-11-25T13:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.114152 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.114212 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.114227 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.114262 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.114273 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:23Z","lastTransitionTime":"2025-11-25T13:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.216483 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.216537 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.216548 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.216570 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.216582 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:23Z","lastTransitionTime":"2025-11-25T13:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.318696 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.318760 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.318769 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.318786 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.318798 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:23Z","lastTransitionTime":"2025-11-25T13:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.422024 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.422096 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.422109 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.422132 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.422147 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:23Z","lastTransitionTime":"2025-11-25T13:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.525321 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.525375 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.525385 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.525402 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.525414 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:23Z","lastTransitionTime":"2025-11-25T13:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.629076 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.629142 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.629170 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.629195 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.629212 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:23Z","lastTransitionTime":"2025-11-25T13:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.732830 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.732910 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.732929 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.732975 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.732997 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:23Z","lastTransitionTime":"2025-11-25T13:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.834946 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.834989 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.835298 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.835318 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.835330 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:23Z","lastTransitionTime":"2025-11-25T13:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.937926 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.937980 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.937989 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.938007 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:23 crc kubenswrapper[4702]: I1125 13:47:23.938019 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:23Z","lastTransitionTime":"2025-11-25T13:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.041139 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.041227 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.041306 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.041330 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.041349 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:24Z","lastTransitionTime":"2025-11-25T13:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.143329 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.143379 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.143391 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.143406 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.143418 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:24Z","lastTransitionTime":"2025-11-25T13:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.246332 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.246394 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.246415 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.246440 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.246458 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:24Z","lastTransitionTime":"2025-11-25T13:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.349689 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.349755 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.349772 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.349798 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.349819 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:24Z","lastTransitionTime":"2025-11-25T13:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.453831 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.453894 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.453917 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.453943 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.453962 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:24Z","lastTransitionTime":"2025-11-25T13:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.557059 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.557098 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.557106 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.557121 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.557132 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:24Z","lastTransitionTime":"2025-11-25T13:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.659899 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.659934 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.659943 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.659957 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.659966 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:24Z","lastTransitionTime":"2025-11-25T13:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.761921 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.761960 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.761971 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.761986 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.761997 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:24Z","lastTransitionTime":"2025-11-25T13:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.800149 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.800277 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.800317 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:24 crc kubenswrapper[4702]: E1125 13:47:24.800470 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:47:24 crc kubenswrapper[4702]: E1125 13:47:24.800569 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:47:24 crc kubenswrapper[4702]: E1125 13:47:24.800665 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fm2cr" podUID="88665994-5305-4de7-8b69-e38b38d5f87c" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.800320 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:47:24 crc kubenswrapper[4702]: E1125 13:47:24.800986 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.864731 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.864795 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.864808 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.864829 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.864846 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:24Z","lastTransitionTime":"2025-11-25T13:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.966967 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.967023 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.967033 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.967046 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:24 crc kubenswrapper[4702]: I1125 13:47:24.967057 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:24Z","lastTransitionTime":"2025-11-25T13:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.069186 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.069265 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.069275 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.069291 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.069301 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:25Z","lastTransitionTime":"2025-11-25T13:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.172526 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.172570 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.172579 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.172595 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.172605 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:25Z","lastTransitionTime":"2025-11-25T13:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.276101 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.276146 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.276156 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.276171 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.276181 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:25Z","lastTransitionTime":"2025-11-25T13:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.379365 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.379414 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.379427 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.379447 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.379466 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:25Z","lastTransitionTime":"2025-11-25T13:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.482080 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.482152 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.482171 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.482225 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.482297 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:25Z","lastTransitionTime":"2025-11-25T13:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.585569 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.585649 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.585664 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.585685 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.585698 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:25Z","lastTransitionTime":"2025-11-25T13:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.688123 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.688169 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.688182 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.688201 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.688216 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:25Z","lastTransitionTime":"2025-11-25T13:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.791267 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.791343 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.791364 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.791394 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.791417 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:25Z","lastTransitionTime":"2025-11-25T13:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.893893 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.893935 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.893943 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.893957 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.893975 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:25Z","lastTransitionTime":"2025-11-25T13:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.997441 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.997488 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.997496 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.997510 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:25 crc kubenswrapper[4702]: I1125 13:47:25.997520 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:25Z","lastTransitionTime":"2025-11-25T13:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.099918 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.099958 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.099966 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.099981 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.099991 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:26Z","lastTransitionTime":"2025-11-25T13:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.201905 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.201946 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.201955 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.201970 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.201981 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:26Z","lastTransitionTime":"2025-11-25T13:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.304656 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.304708 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.304719 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.304741 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.304755 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:26Z","lastTransitionTime":"2025-11-25T13:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.407479 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.407534 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.407547 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.407569 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.407583 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:26Z","lastTransitionTime":"2025-11-25T13:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.505711 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/88665994-5305-4de7-8b69-e38b38d5f87c-metrics-certs\") pod \"network-metrics-daemon-fm2cr\" (UID: \"88665994-5305-4de7-8b69-e38b38d5f87c\") " pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:26 crc kubenswrapper[4702]: E1125 13:47:26.505912 4702 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 13:47:26 crc kubenswrapper[4702]: E1125 13:47:26.506001 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/88665994-5305-4de7-8b69-e38b38d5f87c-metrics-certs podName:88665994-5305-4de7-8b69-e38b38d5f87c nodeName:}" failed. No retries permitted until 2025-11-25 13:47:42.505975938 +0000 UTC m=+71.668626388 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/88665994-5305-4de7-8b69-e38b38d5f87c-metrics-certs") pod "network-metrics-daemon-fm2cr" (UID: "88665994-5305-4de7-8b69-e38b38d5f87c") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.511768 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.511804 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.511812 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.511828 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.511837 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:26Z","lastTransitionTime":"2025-11-25T13:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.613697 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.613741 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.613753 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.613771 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.613783 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:26Z","lastTransitionTime":"2025-11-25T13:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.716019 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.716077 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.716088 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.716102 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.716114 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:26Z","lastTransitionTime":"2025-11-25T13:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.800377 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.800453 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.800420 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:26 crc kubenswrapper[4702]: E1125 13:47:26.800543 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.800572 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:47:26 crc kubenswrapper[4702]: E1125 13:47:26.800661 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fm2cr" podUID="88665994-5305-4de7-8b69-e38b38d5f87c" Nov 25 13:47:26 crc kubenswrapper[4702]: E1125 13:47:26.800757 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:47:26 crc kubenswrapper[4702]: E1125 13:47:26.800914 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.818804 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.818851 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.818859 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.818877 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.818890 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:26Z","lastTransitionTime":"2025-11-25T13:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.922348 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.922419 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.922434 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.922832 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:26 crc kubenswrapper[4702]: I1125 13:47:26.922896 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:26Z","lastTransitionTime":"2025-11-25T13:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.025534 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.025585 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.025594 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.025612 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.025623 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:27Z","lastTransitionTime":"2025-11-25T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.056526 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.068150 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.074691 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.089427 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.102352 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.124706 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a8fa25e9591d484f8dcb56974242924d4f31c7aef86b4376092842e57e1653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d1bfac94a5fe6b382a9a3783250d6758603d00db9ce12ea44a3b54e6018103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c91a144aef48f19006041105594111c382420f9e6556a64580f6952990ef4db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2358ac7b25088c8d1b1cc5648a1fcbfd15211826d4c47f2c98839b9924138b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9649c45bc796d60538bc1edc8f1239c7c5445de7b2f3dbf57dbe7df6ecab7d6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5dfe629efca9398d6dcc541b5315c5b6d8e036cba1c94064883785b2cc7a1a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://427ca923a5d36fe0441dbe5e9dce7e39c374164534025c436e743e2c64a63b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://427ca923a5d36fe0441dbe5e9dce7e39c374164534025c436e743e2c64a63b1c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"message\\\":\\\":134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv in node crc\\\\nF1125 13:47:14.584282 6213 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z]\\\\nI1125 13:47:14.584350 6213 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv after 0 failed attempt(s)\\\\nI1125 13:47:14.584351 6213 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-qqr76 in node crc\\\\nI1125 13:47:14.584357 6213 default_network_controller.go:776] Recording success event on pod\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-p9c5v_openshift-ovn-kubernetes(3c8a3234-01d3-48a9-a5b4-acb46b4218dc)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8700d12c2e236e704fc60136a51489d1d6890a83621480c23a6168d4a0f18f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.130577 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.130647 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.130660 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.130678 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.130690 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:27Z","lastTransitionTime":"2025-11-25T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.138211 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fm2cr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88665994-5305-4de7-8b69-e38b38d5f87c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fm2cr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.155803 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.174041 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.191100 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.208711 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.228122 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60a7f9a29a1e574b4eb0fc72705f8c96b6f613f8a7c164605070077212854cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.233435 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.233485 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.233499 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.233517 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.233530 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:27Z","lastTransitionTime":"2025-11-25T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.257902 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.273914 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.292176 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.305639 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.323702 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.336520 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.336559 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.336568 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.336585 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.336621 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:27Z","lastTransitionTime":"2025-11-25T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.340464 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2efcd45f14690e23b56c19e3282342933f83acd4940f89dd4046b61e4a608b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.356730 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675f4b11-d943-4acc-afde-6f4f45585551\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35b52152fd6304d8c2c961db7d7b144f4c17d8ee855f67e25ca711a89c60a85d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010c83c609ad0554ed60899b53504cd7d859528a4947f45356466d15930c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-djssv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.439803 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.439853 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.439865 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.439881 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.439892 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:27Z","lastTransitionTime":"2025-11-25T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.543694 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.543749 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.543759 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.543779 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.543790 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:27Z","lastTransitionTime":"2025-11-25T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.619988 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.620034 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.620044 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.620062 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.620074 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:27Z","lastTransitionTime":"2025-11-25T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:27 crc kubenswrapper[4702]: E1125 13:47:27.633583 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"96cc93a3-66f3-477f-bbc8-ef3a984ed406\\\",\\\"systemUUID\\\":\\\"bb366f18-cc3c-451f-9d6a-c0cc41864b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.638653 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.638691 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.638706 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.638726 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.638737 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:27Z","lastTransitionTime":"2025-11-25T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:27 crc kubenswrapper[4702]: E1125 13:47:27.650904 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"96cc93a3-66f3-477f-bbc8-ef3a984ed406\\\",\\\"systemUUID\\\":\\\"bb366f18-cc3c-451f-9d6a-c0cc41864b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.655286 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.655323 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.655333 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.655352 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.655364 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:27Z","lastTransitionTime":"2025-11-25T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:27 crc kubenswrapper[4702]: E1125 13:47:27.669768 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"96cc93a3-66f3-477f-bbc8-ef3a984ed406\\\",\\\"systemUUID\\\":\\\"bb366f18-cc3c-451f-9d6a-c0cc41864b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.674393 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.674438 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.674450 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.674474 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.674489 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:27Z","lastTransitionTime":"2025-11-25T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:27 crc kubenswrapper[4702]: E1125 13:47:27.687454 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"96cc93a3-66f3-477f-bbc8-ef3a984ed406\\\",\\\"systemUUID\\\":\\\"bb366f18-cc3c-451f-9d6a-c0cc41864b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.691583 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.691631 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.691645 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.691669 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.691683 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:27Z","lastTransitionTime":"2025-11-25T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:27 crc kubenswrapper[4702]: E1125 13:47:27.703725 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"96cc93a3-66f3-477f-bbc8-ef3a984ed406\\\",\\\"systemUUID\\\":\\\"bb366f18-cc3c-451f-9d6a-c0cc41864b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:27Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:27 crc kubenswrapper[4702]: E1125 13:47:27.703858 4702 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.706040 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.706107 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.706147 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.706171 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.706187 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:27Z","lastTransitionTime":"2025-11-25T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.803264 4702 scope.go:117] "RemoveContainer" containerID="427ca923a5d36fe0441dbe5e9dce7e39c374164534025c436e743e2c64a63b1c" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.809456 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.809504 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.809514 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.809533 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.809545 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:27Z","lastTransitionTime":"2025-11-25T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.912397 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.912879 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.912892 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.912914 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:27 crc kubenswrapper[4702]: I1125 13:47:27.912930 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:27Z","lastTransitionTime":"2025-11-25T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.016323 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.016434 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.016462 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.016491 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.016507 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:28Z","lastTransitionTime":"2025-11-25T13:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.122808 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.122875 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.122890 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.122909 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.122921 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:28Z","lastTransitionTime":"2025-11-25T13:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.189129 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9c5v_3c8a3234-01d3-48a9-a5b4-acb46b4218dc/ovnkube-controller/1.log" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.191895 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" event={"ID":"3c8a3234-01d3-48a9-a5b4-acb46b4218dc","Type":"ContainerStarted","Data":"e042cb58071abb8cf4a711485af66c8fa3259aca650ac4a09cb7d66e893c1cfc"} Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.192542 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.209472 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:28Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.225423 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.225497 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.225510 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.225530 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.225543 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:28Z","lastTransitionTime":"2025-11-25T13:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.231460 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:28Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.248208 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:28Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.270317 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:28Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.285032 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60a7f9a29a1e574b4eb0fc72705f8c96b6f613f8a7c164605070077212854cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:28Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.302325 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:28Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.317295 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:28Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.328262 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.328298 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.328313 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.328327 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.328338 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:28Z","lastTransitionTime":"2025-11-25T13:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.333020 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:28Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.346118 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:28Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.364128 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:28Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.377069 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2efcd45f14690e23b56c19e3282342933f83acd4940f89dd4046b61e4a608b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:28Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.392139 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675f4b11-d943-4acc-afde-6f4f45585551\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35b52152fd6304d8c2c961db7d7b144f4c17d8ee855f67e25ca711a89c60a85d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010c83c609ad0554ed60899b53504cd7d859528a4947f45356466d15930c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-djssv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:28Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.412537 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88cd2da0-a3ca-40cf-9afb-28522fe24930\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff40c37e9756ca143489aafc21e756e213d7204f041b8b807eef932de448210a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b01553a55ccc3121476a5fdfed5532047ecd1e46545aab873da4a7d4718bd96f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b961d759f8b153e4c876e6f1b5eff499fc72a3c828610fc980cba592cac76015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5aaf34113a5a5c1948b672848f8bbcbc0ef69e6a3efe03202c9210016a0c6980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aaf34113a5a5c1948b672848f8bbcbc0ef69e6a3efe03202c9210016a0c6980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:28Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.427909 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:28Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.430925 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.431008 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.431021 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.431063 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.431081 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:28Z","lastTransitionTime":"2025-11-25T13:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.444344 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:28Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.458985 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:28Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.479095 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a8fa25e9591d484f8dcb56974242924d4f31c7aef86b4376092842e57e1653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d1bfac94a5fe6b382a9a3783250d6758603d00db9ce12ea44a3b54e6018103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c91a144aef48f19006041105594111c382420f9e6556a64580f6952990ef4db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2358ac7b25088c8d1b1cc5648a1fcbfd15211826d4c47f2c98839b9924138b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9649c45bc796d60538bc1edc8f1239c7c5445de7b2f3dbf57dbe7df6ecab7d6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5dfe629efca9398d6dcc541b5315c5b6d8e036cba1c94064883785b2cc7a1a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e042cb58071abb8cf4a711485af66c8fa3259aca650ac4a09cb7d66e893c1cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://427ca923a5d36fe0441dbe5e9dce7e39c374164534025c436e743e2c64a63b1c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"message\\\":\\\":134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv in node crc\\\\nF1125 13:47:14.584282 6213 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z]\\\\nI1125 13:47:14.584350 6213 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv after 0 failed attempt(s)\\\\nI1125 13:47:14.584351 6213 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-qqr76 in node crc\\\\nI1125 13:47:14.584357 6213 default_network_controller.go:776] Recording success event on pod\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8700d12c2e236e704fc60136a51489d1d6890a83621480c23a6168d4a0f18f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:28Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.494471 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fm2cr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88665994-5305-4de7-8b69-e38b38d5f87c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fm2cr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:28Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.533445 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.533483 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.533492 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.533507 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.533518 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:28Z","lastTransitionTime":"2025-11-25T13:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.636306 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.636345 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.636355 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.636372 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.636381 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:28Z","lastTransitionTime":"2025-11-25T13:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.739295 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.739336 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.739346 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.739362 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.739372 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:28Z","lastTransitionTime":"2025-11-25T13:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.800096 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.800100 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.800117 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.800275 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:47:28 crc kubenswrapper[4702]: E1125 13:47:28.800402 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fm2cr" podUID="88665994-5305-4de7-8b69-e38b38d5f87c" Nov 25 13:47:28 crc kubenswrapper[4702]: E1125 13:47:28.800459 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:47:28 crc kubenswrapper[4702]: E1125 13:47:28.800537 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:47:28 crc kubenswrapper[4702]: E1125 13:47:28.800600 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.841201 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.841257 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.841267 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.841282 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.841292 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:28Z","lastTransitionTime":"2025-11-25T13:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.943592 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.943650 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.943666 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.943688 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:28 crc kubenswrapper[4702]: I1125 13:47:28.943706 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:28Z","lastTransitionTime":"2025-11-25T13:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.046308 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.046342 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.046351 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.046364 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.046373 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:29Z","lastTransitionTime":"2025-11-25T13:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.148436 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.148472 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.148482 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.148497 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.148508 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:29Z","lastTransitionTime":"2025-11-25T13:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.196738 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9c5v_3c8a3234-01d3-48a9-a5b4-acb46b4218dc/ovnkube-controller/2.log" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.197432 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9c5v_3c8a3234-01d3-48a9-a5b4-acb46b4218dc/ovnkube-controller/1.log" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.201622 4702 generic.go:334] "Generic (PLEG): container finished" podID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerID="e042cb58071abb8cf4a711485af66c8fa3259aca650ac4a09cb7d66e893c1cfc" exitCode=1 Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.201673 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" event={"ID":"3c8a3234-01d3-48a9-a5b4-acb46b4218dc","Type":"ContainerDied","Data":"e042cb58071abb8cf4a711485af66c8fa3259aca650ac4a09cb7d66e893c1cfc"} Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.201714 4702 scope.go:117] "RemoveContainer" containerID="427ca923a5d36fe0441dbe5e9dce7e39c374164534025c436e743e2c64a63b1c" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.202409 4702 scope.go:117] "RemoveContainer" containerID="e042cb58071abb8cf4a711485af66c8fa3259aca650ac4a09cb7d66e893c1cfc" Nov 25 13:47:29 crc kubenswrapper[4702]: E1125 13:47:29.202578 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p9c5v_openshift-ovn-kubernetes(3c8a3234-01d3-48a9-a5b4-acb46b4218dc)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.222415 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:29Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.239200 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60a7f9a29a1e574b4eb0fc72705f8c96b6f613f8a7c164605070077212854cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:29Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.250795 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.250830 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.250839 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.250854 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.250864 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:29Z","lastTransitionTime":"2025-11-25T13:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.252525 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:29Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.264699 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2efcd45f14690e23b56c19e3282342933f83acd4940f89dd4046b61e4a608b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:29Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.282986 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675f4b11-d943-4acc-afde-6f4f45585551\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35b52152fd6304d8c2c961db7d7b144f4c17d8ee855f67e25ca711a89c60a85d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010c83c609ad0554ed60899b53504cd7d859528a4947f45356466d15930c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-djssv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:29Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.304098 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:29Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.317958 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:29Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.331572 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:29Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.342124 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:29Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.353728 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.353791 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.353804 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.353820 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.353831 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:29Z","lastTransitionTime":"2025-11-25T13:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.358390 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:29Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.379667 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a8fa25e9591d484f8dcb56974242924d4f31c7aef86b4376092842e57e1653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d1bfac94a5fe6b382a9a3783250d6758603d00db9ce12ea44a3b54e6018103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c91a144aef48f19006041105594111c382420f9e6556a64580f6952990ef4db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2358ac7b25088c8d1b1cc5648a1fcbfd15211826d4c47f2c98839b9924138b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9649c45bc796d60538bc1edc8f1239c7c5445de7b2f3dbf57dbe7df6ecab7d6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5dfe629efca9398d6dcc541b5315c5b6d8e036cba1c94064883785b2cc7a1a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e042cb58071abb8cf4a711485af66c8fa3259aca650ac4a09cb7d66e893c1cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://427ca923a5d36fe0441dbe5e9dce7e39c374164534025c436e743e2c64a63b1c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"message\\\":\\\":134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv in node crc\\\\nF1125 13:47:14.584282 6213 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:14Z is after 2025-08-24T17:21:41Z]\\\\nI1125 13:47:14.584350 6213 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv after 0 failed attempt(s)\\\\nI1125 13:47:14.584351 6213 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-qqr76 in node crc\\\\nI1125 13:47:14.584357 6213 default_network_controller.go:776] Recording success event on pod\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e042cb58071abb8cf4a711485af66c8fa3259aca650ac4a09cb7d66e893c1cfc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T13:47:28Z\\\",\\\"message\\\":\\\"Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 13:47:28.615062 6389 services_controller.go:445] Built service openshift-image-registry/image-registry LB template configs for network=default: []services.lbConfig(nil)\\\\nF1125 13:47:28.615095 6389 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0x1fcbc00 0x1fcbba0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:28Z is after 2025-08-24T17:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8700d12c2e236e704fc60136a51489d1d6890a83621480c23a6168d4a0f18f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:29Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.391769 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fm2cr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88665994-5305-4de7-8b69-e38b38d5f87c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fm2cr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:29Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.402712 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88cd2da0-a3ca-40cf-9afb-28522fe24930\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff40c37e9756ca143489aafc21e756e213d7204f041b8b807eef932de448210a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b01553a55ccc3121476a5fdfed5532047ecd1e46545aab873da4a7d4718bd96f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b961d759f8b153e4c876e6f1b5eff499fc72a3c828610fc980cba592cac76015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5aaf34113a5a5c1948b672848f8bbcbc0ef69e6a3efe03202c9210016a0c6980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aaf34113a5a5c1948b672848f8bbcbc0ef69e6a3efe03202c9210016a0c6980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:29Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.413318 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:29Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.427327 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:29Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.442338 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:29Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.455153 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:29Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.455731 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.455766 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.455775 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.455789 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.455800 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:29Z","lastTransitionTime":"2025-11-25T13:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.468634 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:29Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.558339 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.558397 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.558406 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.558426 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.558436 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:29Z","lastTransitionTime":"2025-11-25T13:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.660855 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.660911 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.660926 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.660947 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.660958 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:29Z","lastTransitionTime":"2025-11-25T13:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.763560 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.763843 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.763919 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.764021 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.764101 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:29Z","lastTransitionTime":"2025-11-25T13:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.866743 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.866974 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.867077 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.867172 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.867315 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:29Z","lastTransitionTime":"2025-11-25T13:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.970065 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.970674 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.970761 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.970846 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:29 crc kubenswrapper[4702]: I1125 13:47:29.970922 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:29Z","lastTransitionTime":"2025-11-25T13:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.074448 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.074517 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.074533 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.074563 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.074579 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:30Z","lastTransitionTime":"2025-11-25T13:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.176923 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.176977 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.176989 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.177006 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.177018 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:30Z","lastTransitionTime":"2025-11-25T13:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.212034 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9c5v_3c8a3234-01d3-48a9-a5b4-acb46b4218dc/ovnkube-controller/2.log" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.216373 4702 scope.go:117] "RemoveContainer" containerID="e042cb58071abb8cf4a711485af66c8fa3259aca650ac4a09cb7d66e893c1cfc" Nov 25 13:47:30 crc kubenswrapper[4702]: E1125 13:47:30.216673 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p9c5v_openshift-ovn-kubernetes(3c8a3234-01d3-48a9-a5b4-acb46b4218dc)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.231334 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:30Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.247046 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60a7f9a29a1e574b4eb0fc72705f8c96b6f613f8a7c164605070077212854cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:30Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.257473 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:30Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.268993 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:30Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.279467 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.279508 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.279519 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.279535 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.279546 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:30Z","lastTransitionTime":"2025-11-25T13:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.280855 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2efcd45f14690e23b56c19e3282342933f83acd4940f89dd4046b61e4a608b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:30Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.294550 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675f4b11-d943-4acc-afde-6f4f45585551\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35b52152fd6304d8c2c961db7d7b144f4c17d8ee855f67e25ca711a89c60a85d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010c83c609ad0554ed60899b53504cd7d859528a4947f45356466d15930c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-djssv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:30Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.315093 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:30Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.329384 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:30Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.345563 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:30Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.360509 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:30Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.372176 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:30Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.382994 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.383045 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.383055 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.383074 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.383086 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:30Z","lastTransitionTime":"2025-11-25T13:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.393335 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a8fa25e9591d484f8dcb56974242924d4f31c7aef86b4376092842e57e1653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d1bfac94a5fe6b382a9a3783250d6758603d00db9ce12ea44a3b54e6018103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c91a144aef48f19006041105594111c382420f9e6556a64580f6952990ef4db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2358ac7b25088c8d1b1cc5648a1fcbfd15211826d4c47f2c98839b9924138b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9649c45bc796d60538bc1edc8f1239c7c5445de7b2f3dbf57dbe7df6ecab7d6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5dfe629efca9398d6dcc541b5315c5b6d8e036cba1c94064883785b2cc7a1a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e042cb58071abb8cf4a711485af66c8fa3259aca650ac4a09cb7d66e893c1cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e042cb58071abb8cf4a711485af66c8fa3259aca650ac4a09cb7d66e893c1cfc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T13:47:28Z\\\",\\\"message\\\":\\\"Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 13:47:28.615062 6389 services_controller.go:445] Built service openshift-image-registry/image-registry LB template configs for network=default: []services.lbConfig(nil)\\\\nF1125 13:47:28.615095 6389 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0x1fcbc00 0x1fcbba0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:28Z is after 2025-08-24T17:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p9c5v_openshift-ovn-kubernetes(3c8a3234-01d3-48a9-a5b4-acb46b4218dc)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8700d12c2e236e704fc60136a51489d1d6890a83621480c23a6168d4a0f18f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:30Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.406379 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fm2cr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88665994-5305-4de7-8b69-e38b38d5f87c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fm2cr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:30Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.420149 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88cd2da0-a3ca-40cf-9afb-28522fe24930\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff40c37e9756ca143489aafc21e756e213d7204f041b8b807eef932de448210a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b01553a55ccc3121476a5fdfed5532047ecd1e46545aab873da4a7d4718bd96f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b961d759f8b153e4c876e6f1b5eff499fc72a3c828610fc980cba592cac76015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5aaf34113a5a5c1948b672848f8bbcbc0ef69e6a3efe03202c9210016a0c6980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aaf34113a5a5c1948b672848f8bbcbc0ef69e6a3efe03202c9210016a0c6980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:30Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.433891 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:30Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.458931 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:30Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.477940 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:30Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.486658 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.486720 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.486733 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.486754 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.486766 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:30Z","lastTransitionTime":"2025-11-25T13:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.492187 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:30Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.589995 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.590042 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.590054 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.590074 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.590084 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:30Z","lastTransitionTime":"2025-11-25T13:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.693474 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.694508 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.694574 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.694605 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.694625 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:30Z","lastTransitionTime":"2025-11-25T13:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.798871 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.798928 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.798941 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.798964 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.798979 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:30Z","lastTransitionTime":"2025-11-25T13:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.800440 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.800525 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.800574 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:30 crc kubenswrapper[4702]: E1125 13:47:30.800588 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.800689 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:30 crc kubenswrapper[4702]: E1125 13:47:30.800720 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:47:30 crc kubenswrapper[4702]: E1125 13:47:30.800870 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fm2cr" podUID="88665994-5305-4de7-8b69-e38b38d5f87c" Nov 25 13:47:30 crc kubenswrapper[4702]: E1125 13:47:30.800956 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.903199 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.903303 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.903323 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.903355 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:30 crc kubenswrapper[4702]: I1125 13:47:30.903379 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:30Z","lastTransitionTime":"2025-11-25T13:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.007296 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.007354 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.007374 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.007402 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.007426 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:31Z","lastTransitionTime":"2025-11-25T13:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.111194 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.111304 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.111332 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.111369 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.111396 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:31Z","lastTransitionTime":"2025-11-25T13:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.214135 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.214227 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.214286 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.214326 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.214352 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:31Z","lastTransitionTime":"2025-11-25T13:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.317706 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.317753 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.317764 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.317787 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.317803 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:31Z","lastTransitionTime":"2025-11-25T13:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.421021 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.421085 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.421103 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.421135 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.421164 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:31Z","lastTransitionTime":"2025-11-25T13:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.524393 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.524455 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.524471 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.524493 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.524508 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:31Z","lastTransitionTime":"2025-11-25T13:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.627001 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.627540 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.627701 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.627851 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.627986 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:31Z","lastTransitionTime":"2025-11-25T13:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.731471 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.731557 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.731575 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.731602 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.731621 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:31Z","lastTransitionTime":"2025-11-25T13:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.821442 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a8fa25e9591d484f8dcb56974242924d4f31c7aef86b4376092842e57e1653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d1bfac94a5fe6b382a9a3783250d6758603d00db9ce12ea44a3b54e6018103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c91a144aef48f19006041105594111c382420f9e6556a64580f6952990ef4db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2358ac7b25088c8d1b1cc5648a1fcbfd15211826d4c47f2c98839b9924138b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9649c45bc796d60538bc1edc8f1239c7c5445de7b2f3dbf57dbe7df6ecab7d6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5dfe629efca9398d6dcc541b5315c5b6d8e036cba1c94064883785b2cc7a1a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e042cb58071abb8cf4a711485af66c8fa3259aca650ac4a09cb7d66e893c1cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e042cb58071abb8cf4a711485af66c8fa3259aca650ac4a09cb7d66e893c1cfc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T13:47:28Z\\\",\\\"message\\\":\\\"Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 13:47:28.615062 6389 services_controller.go:445] Built service openshift-image-registry/image-registry LB template configs for network=default: []services.lbConfig(nil)\\\\nF1125 13:47:28.615095 6389 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0x1fcbc00 0x1fcbba0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:28Z is after 2025-08-24T17:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p9c5v_openshift-ovn-kubernetes(3c8a3234-01d3-48a9-a5b4-acb46b4218dc)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8700d12c2e236e704fc60136a51489d1d6890a83621480c23a6168d4a0f18f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:31Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.836339 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.836387 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.836397 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.836416 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.836427 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:31Z","lastTransitionTime":"2025-11-25T13:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.838657 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fm2cr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88665994-5305-4de7-8b69-e38b38d5f87c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fm2cr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:31Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.851446 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88cd2da0-a3ca-40cf-9afb-28522fe24930\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff40c37e9756ca143489aafc21e756e213d7204f041b8b807eef932de448210a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b01553a55ccc3121476a5fdfed5532047ecd1e46545aab873da4a7d4718bd96f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b961d759f8b153e4c876e6f1b5eff499fc72a3c828610fc980cba592cac76015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5aaf34113a5a5c1948b672848f8bbcbc0ef69e6a3efe03202c9210016a0c6980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aaf34113a5a5c1948b672848f8bbcbc0ef69e6a3efe03202c9210016a0c6980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:31Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.865021 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:31Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.876418 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:31Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.887269 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:31Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.900702 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:31Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.913127 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:31Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.924366 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:31Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.935677 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:31Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.939723 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.939784 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.939795 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.939811 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.939821 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:31Z","lastTransitionTime":"2025-11-25T13:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.955084 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60a7f9a29a1e574b4eb0fc72705f8c96b6f613f8a7c164605070077212854cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:31Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.969207 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2efcd45f14690e23b56c19e3282342933f83acd4940f89dd4046b61e4a608b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:31Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:31 crc kubenswrapper[4702]: I1125 13:47:31.981338 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675f4b11-d943-4acc-afde-6f4f45585551\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35b52152fd6304d8c2c961db7d7b144f4c17d8ee855f67e25ca711a89c60a85d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010c83c609ad0554ed60899b53504cd7d859528a4947f45356466d15930c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-djssv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:31Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.013704 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.032485 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.043047 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.043144 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.043170 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.043261 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.043287 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:32Z","lastTransitionTime":"2025-11-25T13:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.049987 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.064516 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.079892 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:32Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.145210 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.145509 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.145588 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.145661 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.145756 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:32Z","lastTransitionTime":"2025-11-25T13:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.248131 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.248168 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.248177 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.248190 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.248200 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:32Z","lastTransitionTime":"2025-11-25T13:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.351273 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.351314 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.351326 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.351344 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.351356 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:32Z","lastTransitionTime":"2025-11-25T13:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.454159 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.454444 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.454521 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.454589 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.454652 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:32Z","lastTransitionTime":"2025-11-25T13:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.572421 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.572761 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.572839 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.573003 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.573098 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:32Z","lastTransitionTime":"2025-11-25T13:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.676021 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.676071 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.676080 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.676094 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.676105 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:32Z","lastTransitionTime":"2025-11-25T13:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.778653 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.778696 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.778706 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.778724 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.778735 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:32Z","lastTransitionTime":"2025-11-25T13:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.801097 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.801156 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.801150 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.801122 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:47:32 crc kubenswrapper[4702]: E1125 13:47:32.801276 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:47:32 crc kubenswrapper[4702]: E1125 13:47:32.801429 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:47:32 crc kubenswrapper[4702]: E1125 13:47:32.801476 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fm2cr" podUID="88665994-5305-4de7-8b69-e38b38d5f87c" Nov 25 13:47:32 crc kubenswrapper[4702]: E1125 13:47:32.801545 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.881215 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.881288 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.881301 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.881320 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.881333 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:32Z","lastTransitionTime":"2025-11-25T13:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.984400 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.984442 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.984452 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.984470 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:32 crc kubenswrapper[4702]: I1125 13:47:32.984480 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:32Z","lastTransitionTime":"2025-11-25T13:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.089858 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.089911 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.089920 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.089941 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.089952 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:33Z","lastTransitionTime":"2025-11-25T13:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.192565 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.192615 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.192625 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.192641 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.192651 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:33Z","lastTransitionTime":"2025-11-25T13:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.295664 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.295730 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.295743 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.295761 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.295772 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:33Z","lastTransitionTime":"2025-11-25T13:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.399444 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.399952 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.400057 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.400169 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.400306 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:33Z","lastTransitionTime":"2025-11-25T13:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.503260 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.503314 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.503326 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.503348 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.503361 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:33Z","lastTransitionTime":"2025-11-25T13:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.606645 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.606710 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.606726 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.606749 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.606764 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:33Z","lastTransitionTime":"2025-11-25T13:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.710678 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.710743 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.710756 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.710773 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.710784 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:33Z","lastTransitionTime":"2025-11-25T13:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.813383 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.813420 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.813432 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.813454 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.813466 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:33Z","lastTransitionTime":"2025-11-25T13:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.916389 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.916437 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.916449 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.916498 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:33 crc kubenswrapper[4702]: I1125 13:47:33.916538 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:33Z","lastTransitionTime":"2025-11-25T13:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.019914 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.019995 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.020014 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.020047 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.020068 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:34Z","lastTransitionTime":"2025-11-25T13:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.124908 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.124968 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.124997 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.125021 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.125032 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:34Z","lastTransitionTime":"2025-11-25T13:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.229615 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.229667 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.229677 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.229691 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.229702 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:34Z","lastTransitionTime":"2025-11-25T13:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.332790 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.332844 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.332856 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.332878 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.332892 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:34Z","lastTransitionTime":"2025-11-25T13:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.437223 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.437320 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.437338 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.437370 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.437390 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:34Z","lastTransitionTime":"2025-11-25T13:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.540489 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.540533 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.540541 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.540555 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.540565 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:34Z","lastTransitionTime":"2025-11-25T13:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.643320 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.643641 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.643767 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.643839 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.643906 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:34Z","lastTransitionTime":"2025-11-25T13:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.746535 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.746574 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.746585 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.746598 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.746606 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:34Z","lastTransitionTime":"2025-11-25T13:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.801007 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:34 crc kubenswrapper[4702]: E1125 13:47:34.801374 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.801628 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:47:34 crc kubenswrapper[4702]: E1125 13:47:34.801748 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.802199 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:34 crc kubenswrapper[4702]: E1125 13:47:34.802408 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fm2cr" podUID="88665994-5305-4de7-8b69-e38b38d5f87c" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.802349 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:47:34 crc kubenswrapper[4702]: E1125 13:47:34.802592 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.849027 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.849069 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.849081 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.849097 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.849107 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:34Z","lastTransitionTime":"2025-11-25T13:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.950749 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.951063 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.951218 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.951389 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:34 crc kubenswrapper[4702]: I1125 13:47:34.951531 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:34Z","lastTransitionTime":"2025-11-25T13:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.053972 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.054284 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.054354 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.054418 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.054482 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:35Z","lastTransitionTime":"2025-11-25T13:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.157439 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.157492 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.157504 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.157523 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.157537 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:35Z","lastTransitionTime":"2025-11-25T13:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.260151 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.260668 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.260774 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.260894 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.261006 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:35Z","lastTransitionTime":"2025-11-25T13:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.365105 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.365179 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.365199 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.365251 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.365271 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:35Z","lastTransitionTime":"2025-11-25T13:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.468213 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.468297 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.468308 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.468330 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.468349 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:35Z","lastTransitionTime":"2025-11-25T13:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.571602 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.571657 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.571671 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.571688 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.571699 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:35Z","lastTransitionTime":"2025-11-25T13:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.673789 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.673830 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.673858 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.673872 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.673884 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:35Z","lastTransitionTime":"2025-11-25T13:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.776675 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.776746 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.776760 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.776773 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.776784 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:35Z","lastTransitionTime":"2025-11-25T13:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.879305 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.879372 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.879381 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.879400 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.879411 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:35Z","lastTransitionTime":"2025-11-25T13:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.981952 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.981987 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.981996 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.982011 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:35 crc kubenswrapper[4702]: I1125 13:47:35.982021 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:35Z","lastTransitionTime":"2025-11-25T13:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.084184 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.084218 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.084225 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.084265 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.084278 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:36Z","lastTransitionTime":"2025-11-25T13:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.186707 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.186746 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.186757 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.186772 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.186783 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:36Z","lastTransitionTime":"2025-11-25T13:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.288386 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.288424 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.288435 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.288452 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.288462 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:36Z","lastTransitionTime":"2025-11-25T13:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.390921 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.390983 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.390993 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.391008 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.391037 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:36Z","lastTransitionTime":"2025-11-25T13:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.493137 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.493172 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.493180 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.493194 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.493204 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:36Z","lastTransitionTime":"2025-11-25T13:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.595573 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.595614 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.595623 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.595638 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.595650 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:36Z","lastTransitionTime":"2025-11-25T13:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.698875 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.698914 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.698923 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.698939 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.698949 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:36Z","lastTransitionTime":"2025-11-25T13:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.799986 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.800019 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:36 crc kubenswrapper[4702]: E1125 13:47:36.800246 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fm2cr" podUID="88665994-5305-4de7-8b69-e38b38d5f87c" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.800303 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.800331 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:47:36 crc kubenswrapper[4702]: E1125 13:47:36.800395 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:47:36 crc kubenswrapper[4702]: E1125 13:47:36.800547 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.800677 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.800736 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.800749 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:36 crc kubenswrapper[4702]: E1125 13:47:36.800748 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.800771 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.800781 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:36Z","lastTransitionTime":"2025-11-25T13:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.902907 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.902951 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.902961 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.902976 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:36 crc kubenswrapper[4702]: I1125 13:47:36.902985 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:36Z","lastTransitionTime":"2025-11-25T13:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.004999 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.005046 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.005058 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.005077 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.005089 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:37Z","lastTransitionTime":"2025-11-25T13:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.106919 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.106957 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.106967 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.106980 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.106989 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:37Z","lastTransitionTime":"2025-11-25T13:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.209229 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.209317 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.209330 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.209348 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.209361 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:37Z","lastTransitionTime":"2025-11-25T13:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.312528 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.312580 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.312596 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.312618 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.312635 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:37Z","lastTransitionTime":"2025-11-25T13:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.415831 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.415885 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.415897 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.415916 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.415929 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:37Z","lastTransitionTime":"2025-11-25T13:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.518590 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.518637 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.518649 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.518668 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.518681 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:37Z","lastTransitionTime":"2025-11-25T13:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.621542 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.621578 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.621591 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.621608 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.621622 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:37Z","lastTransitionTime":"2025-11-25T13:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.723574 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.723616 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.723625 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.723639 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.723649 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:37Z","lastTransitionTime":"2025-11-25T13:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.826427 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.826503 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.826526 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.826556 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.826577 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:37Z","lastTransitionTime":"2025-11-25T13:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.928625 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.928669 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.928684 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.928701 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.928713 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:37Z","lastTransitionTime":"2025-11-25T13:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.999467 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.999526 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.999541 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.999559 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:37 crc kubenswrapper[4702]: I1125 13:47:37.999571 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:37Z","lastTransitionTime":"2025-11-25T13:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:38 crc kubenswrapper[4702]: E1125 13:47:38.017598 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"96cc93a3-66f3-477f-bbc8-ef3a984ed406\\\",\\\"systemUUID\\\":\\\"bb366f18-cc3c-451f-9d6a-c0cc41864b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:38Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.021459 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.021498 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.021508 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.021546 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.021557 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:38Z","lastTransitionTime":"2025-11-25T13:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:38 crc kubenswrapper[4702]: E1125 13:47:38.039689 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"96cc93a3-66f3-477f-bbc8-ef3a984ed406\\\",\\\"systemUUID\\\":\\\"bb366f18-cc3c-451f-9d6a-c0cc41864b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:38Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.043533 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.043592 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.043602 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.043619 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.043631 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:38Z","lastTransitionTime":"2025-11-25T13:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:38 crc kubenswrapper[4702]: E1125 13:47:38.055693 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"96cc93a3-66f3-477f-bbc8-ef3a984ed406\\\",\\\"systemUUID\\\":\\\"bb366f18-cc3c-451f-9d6a-c0cc41864b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:38Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.059583 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.059624 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.059639 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.059657 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.059674 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:38Z","lastTransitionTime":"2025-11-25T13:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:38 crc kubenswrapper[4702]: E1125 13:47:38.070915 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"96cc93a3-66f3-477f-bbc8-ef3a984ed406\\\",\\\"systemUUID\\\":\\\"bb366f18-cc3c-451f-9d6a-c0cc41864b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:38Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.074809 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.074843 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.074855 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.074875 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.074888 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:38Z","lastTransitionTime":"2025-11-25T13:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:38 crc kubenswrapper[4702]: E1125 13:47:38.088182 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"96cc93a3-66f3-477f-bbc8-ef3a984ed406\\\",\\\"systemUUID\\\":\\\"bb366f18-cc3c-451f-9d6a-c0cc41864b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:38Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:38 crc kubenswrapper[4702]: E1125 13:47:38.088313 4702 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.089857 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.089898 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.089908 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.089923 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.089934 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:38Z","lastTransitionTime":"2025-11-25T13:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.192731 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.192801 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.192823 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.192852 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.192875 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:38Z","lastTransitionTime":"2025-11-25T13:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.295884 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.295939 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.295951 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.295971 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.295985 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:38Z","lastTransitionTime":"2025-11-25T13:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.398400 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.398462 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.398479 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.398505 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.398518 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:38Z","lastTransitionTime":"2025-11-25T13:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.501036 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.501086 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.501096 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.501116 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.501126 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:38Z","lastTransitionTime":"2025-11-25T13:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.604084 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.604135 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.604145 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.604164 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.604179 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:38Z","lastTransitionTime":"2025-11-25T13:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.706790 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.706842 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.706855 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.706874 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.706888 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:38Z","lastTransitionTime":"2025-11-25T13:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.800430 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.800465 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.800527 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:38 crc kubenswrapper[4702]: E1125 13:47:38.800553 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:47:38 crc kubenswrapper[4702]: E1125 13:47:38.800637 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.800704 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:47:38 crc kubenswrapper[4702]: E1125 13:47:38.800786 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fm2cr" podUID="88665994-5305-4de7-8b69-e38b38d5f87c" Nov 25 13:47:38 crc kubenswrapper[4702]: E1125 13:47:38.800934 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.809516 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.809558 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.809568 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.809587 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.809601 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:38Z","lastTransitionTime":"2025-11-25T13:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.921288 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.921354 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.921373 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.921397 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:38 crc kubenswrapper[4702]: I1125 13:47:38.921413 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:38Z","lastTransitionTime":"2025-11-25T13:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.023999 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.024063 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.024074 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.024100 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.024122 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:39Z","lastTransitionTime":"2025-11-25T13:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.127153 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.127223 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.127263 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.127289 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.127303 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:39Z","lastTransitionTime":"2025-11-25T13:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.231106 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.231158 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.231170 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.231187 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.231198 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:39Z","lastTransitionTime":"2025-11-25T13:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.333713 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.333764 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.333780 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.333798 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.333809 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:39Z","lastTransitionTime":"2025-11-25T13:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.436465 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.436536 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.436552 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.436579 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.436593 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:39Z","lastTransitionTime":"2025-11-25T13:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.539259 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.539317 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.539332 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.539350 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.539362 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:39Z","lastTransitionTime":"2025-11-25T13:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.642920 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.643002 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.643019 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.643050 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.643076 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:39Z","lastTransitionTime":"2025-11-25T13:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.745980 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.746036 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.746050 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.746070 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.746084 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:39Z","lastTransitionTime":"2025-11-25T13:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.848686 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.848747 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.848761 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.848790 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.848806 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:39Z","lastTransitionTime":"2025-11-25T13:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.952122 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.952163 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.952174 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.952195 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:39 crc kubenswrapper[4702]: I1125 13:47:39.952207 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:39Z","lastTransitionTime":"2025-11-25T13:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.055046 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.055115 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.055134 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.055218 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.055267 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:40Z","lastTransitionTime":"2025-11-25T13:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.157997 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.158100 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.158115 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.158137 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.158150 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:40Z","lastTransitionTime":"2025-11-25T13:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.260731 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.260789 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.260806 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.260829 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.260846 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:40Z","lastTransitionTime":"2025-11-25T13:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.363459 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.363523 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.363536 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.363562 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.363578 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:40Z","lastTransitionTime":"2025-11-25T13:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.466391 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.466440 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.466450 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.466477 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.466493 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:40Z","lastTransitionTime":"2025-11-25T13:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.569310 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.569357 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.569372 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.569393 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.569407 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:40Z","lastTransitionTime":"2025-11-25T13:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.672624 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.672675 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.672685 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.672704 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.672715 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:40Z","lastTransitionTime":"2025-11-25T13:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.775329 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.775407 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.775430 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.775460 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.775485 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:40Z","lastTransitionTime":"2025-11-25T13:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.800435 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.800559 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:40 crc kubenswrapper[4702]: E1125 13:47:40.800705 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.800735 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.800798 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:47:40 crc kubenswrapper[4702]: E1125 13:47:40.801009 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:47:40 crc kubenswrapper[4702]: E1125 13:47:40.801063 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:47:40 crc kubenswrapper[4702]: E1125 13:47:40.801137 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fm2cr" podUID="88665994-5305-4de7-8b69-e38b38d5f87c" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.879212 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.879298 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.879311 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.879330 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.879339 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:40Z","lastTransitionTime":"2025-11-25T13:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.982322 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.982382 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.982397 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.982419 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:40 crc kubenswrapper[4702]: I1125 13:47:40.982431 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:40Z","lastTransitionTime":"2025-11-25T13:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.085532 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.085576 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.085587 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.085604 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.085618 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:41Z","lastTransitionTime":"2025-11-25T13:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.188428 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.188471 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.188509 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.188525 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.188539 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:41Z","lastTransitionTime":"2025-11-25T13:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.291100 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.291146 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.291159 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.291175 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.291189 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:41Z","lastTransitionTime":"2025-11-25T13:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.393897 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.393940 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.393953 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.393969 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.393981 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:41Z","lastTransitionTime":"2025-11-25T13:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.497555 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.497636 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.497658 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.497694 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.497720 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:41Z","lastTransitionTime":"2025-11-25T13:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.601092 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.601145 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.601157 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.601177 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.601190 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:41Z","lastTransitionTime":"2025-11-25T13:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.704014 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.704069 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.704088 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.704109 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.704120 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:41Z","lastTransitionTime":"2025-11-25T13:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.808469 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.808505 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.808515 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.808531 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.808542 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:41Z","lastTransitionTime":"2025-11-25T13:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.818671 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:41Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.833598 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60a7f9a29a1e574b4eb0fc72705f8c96b6f613f8a7c164605070077212854cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:41Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.845035 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675f4b11-d943-4acc-afde-6f4f45585551\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35b52152fd6304d8c2c961db7d7b144f4c17d8ee855f67e25ca711a89c60a85d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010c83c609ad0554ed60899b53504cd7d859528a4947f45356466d15930c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-djssv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:41Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.865956 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:41Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.880923 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:41Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.892582 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:41Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.901495 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:41Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.910314 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.910342 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.910353 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.910369 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.910380 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:41Z","lastTransitionTime":"2025-11-25T13:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.920383 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:41Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.931027 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2efcd45f14690e23b56c19e3282342933f83acd4940f89dd4046b61e4a608b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:41Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.946332 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fm2cr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88665994-5305-4de7-8b69-e38b38d5f87c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fm2cr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:41Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.959098 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88cd2da0-a3ca-40cf-9afb-28522fe24930\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff40c37e9756ca143489aafc21e756e213d7204f041b8b807eef932de448210a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b01553a55ccc3121476a5fdfed5532047ecd1e46545aab873da4a7d4718bd96f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b961d759f8b153e4c876e6f1b5eff499fc72a3c828610fc980cba592cac76015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5aaf34113a5a5c1948b672848f8bbcbc0ef69e6a3efe03202c9210016a0c6980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aaf34113a5a5c1948b672848f8bbcbc0ef69e6a3efe03202c9210016a0c6980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:41Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.971408 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:41Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.983382 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:41Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:41 crc kubenswrapper[4702]: I1125 13:47:41.994359 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:41Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.011735 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.011773 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.011787 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.011805 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.011816 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:42Z","lastTransitionTime":"2025-11-25T13:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.013742 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a8fa25e9591d484f8dcb56974242924d4f31c7aef86b4376092842e57e1653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d1bfac94a5fe6b382a9a3783250d6758603d00db9ce12ea44a3b54e6018103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c91a144aef48f19006041105594111c382420f9e6556a64580f6952990ef4db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2358ac7b25088c8d1b1cc5648a1fcbfd15211826d4c47f2c98839b9924138b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9649c45bc796d60538bc1edc8f1239c7c5445de7b2f3dbf57dbe7df6ecab7d6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5dfe629efca9398d6dcc541b5315c5b6d8e036cba1c94064883785b2cc7a1a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e042cb58071abb8cf4a711485af66c8fa3259aca650ac4a09cb7d66e893c1cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e042cb58071abb8cf4a711485af66c8fa3259aca650ac4a09cb7d66e893c1cfc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T13:47:28Z\\\",\\\"message\\\":\\\"Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 13:47:28.615062 6389 services_controller.go:445] Built service openshift-image-registry/image-registry LB template configs for network=default: []services.lbConfig(nil)\\\\nF1125 13:47:28.615095 6389 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0x1fcbc00 0x1fcbba0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:28Z is after 2025-08-24T17:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p9c5v_openshift-ovn-kubernetes(3c8a3234-01d3-48a9-a5b4-acb46b4218dc)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8700d12c2e236e704fc60136a51489d1d6890a83621480c23a6168d4a0f18f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:42Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.026819 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:42Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.040010 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:42Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.050290 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:42Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.114271 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.114313 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.114323 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.114342 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.114352 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:42Z","lastTransitionTime":"2025-11-25T13:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.217117 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.217498 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.217543 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.217576 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.217595 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:42Z","lastTransitionTime":"2025-11-25T13:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.320739 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.321307 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.321411 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.321516 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.321609 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:42Z","lastTransitionTime":"2025-11-25T13:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.424820 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.424879 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.424888 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.424906 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.424918 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:42Z","lastTransitionTime":"2025-11-25T13:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.509954 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/88665994-5305-4de7-8b69-e38b38d5f87c-metrics-certs\") pod \"network-metrics-daemon-fm2cr\" (UID: \"88665994-5305-4de7-8b69-e38b38d5f87c\") " pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:42 crc kubenswrapper[4702]: E1125 13:47:42.510170 4702 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 13:47:42 crc kubenswrapper[4702]: E1125 13:47:42.510586 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/88665994-5305-4de7-8b69-e38b38d5f87c-metrics-certs podName:88665994-5305-4de7-8b69-e38b38d5f87c nodeName:}" failed. No retries permitted until 2025-11-25 13:48:14.510556479 +0000 UTC m=+103.673206929 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/88665994-5305-4de7-8b69-e38b38d5f87c-metrics-certs") pod "network-metrics-daemon-fm2cr" (UID: "88665994-5305-4de7-8b69-e38b38d5f87c") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.527923 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.527969 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.527977 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.527993 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.528009 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:42Z","lastTransitionTime":"2025-11-25T13:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.630718 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.630958 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.631064 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.631152 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.631224 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:42Z","lastTransitionTime":"2025-11-25T13:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.733935 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.733994 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.734005 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.734024 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.734033 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:42Z","lastTransitionTime":"2025-11-25T13:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.801031 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.801094 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:42 crc kubenswrapper[4702]: E1125 13:47:42.801135 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:47:42 crc kubenswrapper[4702]: E1125 13:47:42.801179 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fm2cr" podUID="88665994-5305-4de7-8b69-e38b38d5f87c" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.801214 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:42 crc kubenswrapper[4702]: E1125 13:47:42.801435 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.801049 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:47:42 crc kubenswrapper[4702]: E1125 13:47:42.802022 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.836845 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.836886 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.836898 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.836912 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.836924 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:42Z","lastTransitionTime":"2025-11-25T13:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.939295 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.939537 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.939654 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.939754 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:42 crc kubenswrapper[4702]: I1125 13:47:42.939843 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:42Z","lastTransitionTime":"2025-11-25T13:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.042586 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.042625 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.042636 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.042653 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.042687 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:43Z","lastTransitionTime":"2025-11-25T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.145256 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.145314 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.145323 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.145339 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.145350 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:43Z","lastTransitionTime":"2025-11-25T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.247561 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.247602 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.247613 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.247629 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.247639 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:43Z","lastTransitionTime":"2025-11-25T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.349863 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.349906 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.349916 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.349931 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.349940 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:43Z","lastTransitionTime":"2025-11-25T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.452084 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.452121 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.452129 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.452145 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.452158 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:43Z","lastTransitionTime":"2025-11-25T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.555052 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.555099 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.555115 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.555134 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.555147 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:43Z","lastTransitionTime":"2025-11-25T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.657100 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.657136 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.657144 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.657159 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.657168 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:43Z","lastTransitionTime":"2025-11-25T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.759649 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.759693 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.759704 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.759717 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.759727 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:43Z","lastTransitionTime":"2025-11-25T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.861905 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.861964 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.861980 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.862000 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.862011 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:43Z","lastTransitionTime":"2025-11-25T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.963879 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.963918 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.963927 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.963942 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:43 crc kubenswrapper[4702]: I1125 13:47:43.963954 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:43Z","lastTransitionTime":"2025-11-25T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.066341 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.066469 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.066484 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.066500 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.066511 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:44Z","lastTransitionTime":"2025-11-25T13:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.168931 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.168973 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.168984 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.169000 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.169012 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:44Z","lastTransitionTime":"2025-11-25T13:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.271268 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.271302 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.271312 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.271327 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.271340 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:44Z","lastTransitionTime":"2025-11-25T13:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.373660 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.373713 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.373725 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.373741 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.373753 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:44Z","lastTransitionTime":"2025-11-25T13:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.476263 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.476324 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.476338 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.476356 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.476368 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:44Z","lastTransitionTime":"2025-11-25T13:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.579099 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.579130 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.579139 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.579153 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.579163 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:44Z","lastTransitionTime":"2025-11-25T13:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.681404 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.681431 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.681439 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.681454 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.681463 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:44Z","lastTransitionTime":"2025-11-25T13:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.784450 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.784483 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.784492 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.784506 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.784515 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:44Z","lastTransitionTime":"2025-11-25T13:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.800873 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.800917 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:44 crc kubenswrapper[4702]: E1125 13:47:44.801000 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.801096 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.801143 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:47:44 crc kubenswrapper[4702]: E1125 13:47:44.801398 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:47:44 crc kubenswrapper[4702]: E1125 13:47:44.801433 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:47:44 crc kubenswrapper[4702]: E1125 13:47:44.801855 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fm2cr" podUID="88665994-5305-4de7-8b69-e38b38d5f87c" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.802251 4702 scope.go:117] "RemoveContainer" containerID="e042cb58071abb8cf4a711485af66c8fa3259aca650ac4a09cb7d66e893c1cfc" Nov 25 13:47:44 crc kubenswrapper[4702]: E1125 13:47:44.802610 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p9c5v_openshift-ovn-kubernetes(3c8a3234-01d3-48a9-a5b4-acb46b4218dc)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.887668 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.887959 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.887969 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.887984 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.887994 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:44Z","lastTransitionTime":"2025-11-25T13:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.990524 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.990567 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.990578 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.990596 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:44 crc kubenswrapper[4702]: I1125 13:47:44.990607 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:44Z","lastTransitionTime":"2025-11-25T13:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.093643 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.094224 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.094287 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.094306 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.094317 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:45Z","lastTransitionTime":"2025-11-25T13:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.195878 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.195910 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.195919 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.195934 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.195942 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:45Z","lastTransitionTime":"2025-11-25T13:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.298530 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.298566 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.298574 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.298588 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.298598 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:45Z","lastTransitionTime":"2025-11-25T13:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.402372 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.402428 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.402438 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.402455 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.402466 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:45Z","lastTransitionTime":"2025-11-25T13:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.506521 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.506612 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.506633 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.506667 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.506689 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:45Z","lastTransitionTime":"2025-11-25T13:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.608972 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.609008 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.609015 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.609028 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.609037 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:45Z","lastTransitionTime":"2025-11-25T13:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.713046 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.713101 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.713117 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.713136 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.713154 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:45Z","lastTransitionTime":"2025-11-25T13:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.815975 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.816017 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.816032 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.816047 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.816056 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:45Z","lastTransitionTime":"2025-11-25T13:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.918048 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.918081 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.918089 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.918102 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:45 crc kubenswrapper[4702]: I1125 13:47:45.918112 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:45Z","lastTransitionTime":"2025-11-25T13:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.020119 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.020161 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.020169 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.020185 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.020195 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:46Z","lastTransitionTime":"2025-11-25T13:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.122915 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.122962 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.122971 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.122986 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.122996 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:46Z","lastTransitionTime":"2025-11-25T13:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.224973 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.225012 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.225020 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.225036 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.225044 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:46Z","lastTransitionTime":"2025-11-25T13:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.328102 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.328161 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.328177 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.328201 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.328218 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:46Z","lastTransitionTime":"2025-11-25T13:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.430342 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.430378 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.430386 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.430403 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.430412 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:46Z","lastTransitionTime":"2025-11-25T13:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.532517 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.532548 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.532556 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.532570 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.532579 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:46Z","lastTransitionTime":"2025-11-25T13:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.635225 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.635286 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.635295 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.635308 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.635317 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:46Z","lastTransitionTime":"2025-11-25T13:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.737112 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.737143 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.737150 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.737163 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.737171 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:46Z","lastTransitionTime":"2025-11-25T13:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.800633 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.800683 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.800740 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:47:46 crc kubenswrapper[4702]: E1125 13:47:46.800859 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.800895 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:46 crc kubenswrapper[4702]: E1125 13:47:46.801019 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:47:46 crc kubenswrapper[4702]: E1125 13:47:46.801051 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:47:46 crc kubenswrapper[4702]: E1125 13:47:46.801138 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fm2cr" podUID="88665994-5305-4de7-8b69-e38b38d5f87c" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.839504 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.839547 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.839558 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.839576 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.839588 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:46Z","lastTransitionTime":"2025-11-25T13:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.942227 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.942303 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.942315 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.942330 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:46 crc kubenswrapper[4702]: I1125 13:47:46.942339 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:46Z","lastTransitionTime":"2025-11-25T13:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.045139 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.045268 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.045297 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.045332 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.045363 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:47Z","lastTransitionTime":"2025-11-25T13:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.147908 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.147941 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.147950 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.147962 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.147971 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:47Z","lastTransitionTime":"2025-11-25T13:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.251110 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.251157 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.251168 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.251183 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.251193 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:47Z","lastTransitionTime":"2025-11-25T13:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.276563 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gd2xs_9e735a90-cf24-47de-b0e6-57d27a971b21/kube-multus/0.log" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.276605 4702 generic.go:334] "Generic (PLEG): container finished" podID="9e735a90-cf24-47de-b0e6-57d27a971b21" containerID="b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694" exitCode=1 Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.276628 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gd2xs" event={"ID":"9e735a90-cf24-47de-b0e6-57d27a971b21","Type":"ContainerDied","Data":"b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694"} Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.276902 4702 scope.go:117] "RemoveContainer" containerID="b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.313273 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.325952 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.339195 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.349107 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.353043 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.353105 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.353127 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.353153 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.353171 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:47Z","lastTransitionTime":"2025-11-25T13:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.360508 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T13:47:46Z\\\",\\\"message\\\":\\\"2025-11-25T13:47:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_48eac4c0-a560-4b00-86f2-65d9e25b2f5b\\\\n2025-11-25T13:47:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_48eac4c0-a560-4b00-86f2-65d9e25b2f5b to /host/opt/cni/bin/\\\\n2025-11-25T13:47:01Z [verbose] multus-daemon started\\\\n2025-11-25T13:47:01Z [verbose] Readiness Indicator file check\\\\n2025-11-25T13:47:46Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.372108 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2efcd45f14690e23b56c19e3282342933f83acd4940f89dd4046b61e4a608b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.384156 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675f4b11-d943-4acc-afde-6f4f45585551\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35b52152fd6304d8c2c961db7d7b144f4c17d8ee855f67e25ca711a89c60a85d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010c83c609ad0554ed60899b53504cd7d859528a4947f45356466d15930c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-djssv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.395537 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88cd2da0-a3ca-40cf-9afb-28522fe24930\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff40c37e9756ca143489aafc21e756e213d7204f041b8b807eef932de448210a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b01553a55ccc3121476a5fdfed5532047ecd1e46545aab873da4a7d4718bd96f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b961d759f8b153e4c876e6f1b5eff499fc72a3c828610fc980cba592cac76015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5aaf34113a5a5c1948b672848f8bbcbc0ef69e6a3efe03202c9210016a0c6980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aaf34113a5a5c1948b672848f8bbcbc0ef69e6a3efe03202c9210016a0c6980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.408484 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.418955 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.430368 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.447664 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a8fa25e9591d484f8dcb56974242924d4f31c7aef86b4376092842e57e1653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d1bfac94a5fe6b382a9a3783250d6758603d00db9ce12ea44a3b54e6018103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c91a144aef48f19006041105594111c382420f9e6556a64580f6952990ef4db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2358ac7b25088c8d1b1cc5648a1fcbfd15211826d4c47f2c98839b9924138b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9649c45bc796d60538bc1edc8f1239c7c5445de7b2f3dbf57dbe7df6ecab7d6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5dfe629efca9398d6dcc541b5315c5b6d8e036cba1c94064883785b2cc7a1a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e042cb58071abb8cf4a711485af66c8fa3259aca650ac4a09cb7d66e893c1cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e042cb58071abb8cf4a711485af66c8fa3259aca650ac4a09cb7d66e893c1cfc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T13:47:28Z\\\",\\\"message\\\":\\\"Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 13:47:28.615062 6389 services_controller.go:445] Built service openshift-image-registry/image-registry LB template configs for network=default: []services.lbConfig(nil)\\\\nF1125 13:47:28.615095 6389 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0x1fcbc00 0x1fcbba0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:28Z is after 2025-08-24T17:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p9c5v_openshift-ovn-kubernetes(3c8a3234-01d3-48a9-a5b4-acb46b4218dc)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8700d12c2e236e704fc60136a51489d1d6890a83621480c23a6168d4a0f18f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.456407 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.456447 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.456460 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.456476 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.456487 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:47Z","lastTransitionTime":"2025-11-25T13:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.458544 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fm2cr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88665994-5305-4de7-8b69-e38b38d5f87c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fm2cr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.470918 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.482955 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.495449 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.506845 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.525424 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60a7f9a29a1e574b4eb0fc72705f8c96b6f613f8a7c164605070077212854cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.559267 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.559308 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.559357 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.559377 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.559388 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:47Z","lastTransitionTime":"2025-11-25T13:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.662205 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.662302 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.662317 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.662336 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.662348 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:47Z","lastTransitionTime":"2025-11-25T13:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.765070 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.765162 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.765189 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.765218 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.765279 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:47Z","lastTransitionTime":"2025-11-25T13:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.814325 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.868290 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.868330 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.868339 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.868351 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.868360 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:47Z","lastTransitionTime":"2025-11-25T13:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.970698 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.970741 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.970753 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.970770 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:47 crc kubenswrapper[4702]: I1125 13:47:47.970781 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:47Z","lastTransitionTime":"2025-11-25T13:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.072415 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.072458 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.072467 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.072481 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.072492 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:48Z","lastTransitionTime":"2025-11-25T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.094307 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.094343 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.094353 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.094368 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.094378 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:48Z","lastTransitionTime":"2025-11-25T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:48 crc kubenswrapper[4702]: E1125 13:47:48.106005 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"96cc93a3-66f3-477f-bbc8-ef3a984ed406\\\",\\\"systemUUID\\\":\\\"bb366f18-cc3c-451f-9d6a-c0cc41864b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.111538 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.111571 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.111581 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.111600 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.111611 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:48Z","lastTransitionTime":"2025-11-25T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:48 crc kubenswrapper[4702]: E1125 13:47:48.124346 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"96cc93a3-66f3-477f-bbc8-ef3a984ed406\\\",\\\"systemUUID\\\":\\\"bb366f18-cc3c-451f-9d6a-c0cc41864b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.128521 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.128601 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.128694 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.128804 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.128911 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:48Z","lastTransitionTime":"2025-11-25T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:48 crc kubenswrapper[4702]: E1125 13:47:48.148871 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"96cc93a3-66f3-477f-bbc8-ef3a984ed406\\\",\\\"systemUUID\\\":\\\"bb366f18-cc3c-451f-9d6a-c0cc41864b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.151907 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.151941 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.151950 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.151966 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.151977 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:48Z","lastTransitionTime":"2025-11-25T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:48 crc kubenswrapper[4702]: E1125 13:47:48.165706 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"96cc93a3-66f3-477f-bbc8-ef3a984ed406\\\",\\\"systemUUID\\\":\\\"bb366f18-cc3c-451f-9d6a-c0cc41864b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.169656 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.169819 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.169864 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.170031 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.170128 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:48Z","lastTransitionTime":"2025-11-25T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:48 crc kubenswrapper[4702]: E1125 13:47:48.191354 4702 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"96cc93a3-66f3-477f-bbc8-ef3a984ed406\\\",\\\"systemUUID\\\":\\\"bb366f18-cc3c-451f-9d6a-c0cc41864b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:48 crc kubenswrapper[4702]: E1125 13:47:48.191512 4702 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.193110 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.193147 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.193157 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.193172 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.193183 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:48Z","lastTransitionTime":"2025-11-25T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.282001 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gd2xs_9e735a90-cf24-47de-b0e6-57d27a971b21/kube-multus/0.log" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.282453 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gd2xs" event={"ID":"9e735a90-cf24-47de-b0e6-57d27a971b21","Type":"ContainerStarted","Data":"10ef8fba2e92332ed66f05f36bc57004f1d12c53689621106a3606f71d8aee9b"} Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.301453 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.301500 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.301509 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.301563 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.301574 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:48Z","lastTransitionTime":"2025-11-25T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.306689 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.322835 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.339883 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10ef8fba2e92332ed66f05f36bc57004f1d12c53689621106a3606f71d8aee9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T13:47:46Z\\\",\\\"message\\\":\\\"2025-11-25T13:47:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_48eac4c0-a560-4b00-86f2-65d9e25b2f5b\\\\n2025-11-25T13:47:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_48eac4c0-a560-4b00-86f2-65d9e25b2f5b to /host/opt/cni/bin/\\\\n2025-11-25T13:47:01Z [verbose] multus-daemon started\\\\n2025-11-25T13:47:01Z [verbose] Readiness Indicator file check\\\\n2025-11-25T13:47:46Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.352213 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2efcd45f14690e23b56c19e3282342933f83acd4940f89dd4046b61e4a608b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.365340 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675f4b11-d943-4acc-afde-6f4f45585551\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35b52152fd6304d8c2c961db7d7b144f4c17d8ee855f67e25ca711a89c60a85d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010c83c609ad0554ed60899b53504cd7d859528a4947f45356466d15930c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-djssv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.394080 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.406593 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.406629 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.406640 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.406656 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.406669 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:48Z","lastTransitionTime":"2025-11-25T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.420309 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.442127 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.458431 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.474547 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.507168 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a8fa25e9591d484f8dcb56974242924d4f31c7aef86b4376092842e57e1653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d1bfac94a5fe6b382a9a3783250d6758603d00db9ce12ea44a3b54e6018103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c91a144aef48f19006041105594111c382420f9e6556a64580f6952990ef4db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2358ac7b25088c8d1b1cc5648a1fcbfd15211826d4c47f2c98839b9924138b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9649c45bc796d60538bc1edc8f1239c7c5445de7b2f3dbf57dbe7df6ecab7d6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5dfe629efca9398d6dcc541b5315c5b6d8e036cba1c94064883785b2cc7a1a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e042cb58071abb8cf4a711485af66c8fa3259aca650ac4a09cb7d66e893c1cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e042cb58071abb8cf4a711485af66c8fa3259aca650ac4a09cb7d66e893c1cfc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T13:47:28Z\\\",\\\"message\\\":\\\"Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 13:47:28.615062 6389 services_controller.go:445] Built service openshift-image-registry/image-registry LB template configs for network=default: []services.lbConfig(nil)\\\\nF1125 13:47:28.615095 6389 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0x1fcbc00 0x1fcbba0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:28Z is after 2025-08-24T17:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p9c5v_openshift-ovn-kubernetes(3c8a3234-01d3-48a9-a5b4-acb46b4218dc)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8700d12c2e236e704fc60136a51489d1d6890a83621480c23a6168d4a0f18f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.509414 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.509452 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.509467 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.509484 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.509497 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:48Z","lastTransitionTime":"2025-11-25T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.523685 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fm2cr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88665994-5305-4de7-8b69-e38b38d5f87c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fm2cr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.537093 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0f51e5a-21b4-4b12-884d-823946038ffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90db72f00d2ebd6a338e6ec5396349087083c2ad55dd120e712b1e4916b7c5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e08782f884652e08e363cd5cad0db72b87dedb7eaefa8db537a43ebdfc778325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e08782f884652e08e363cd5cad0db72b87dedb7eaefa8db537a43ebdfc778325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.553366 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88cd2da0-a3ca-40cf-9afb-28522fe24930\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff40c37e9756ca143489aafc21e756e213d7204f041b8b807eef932de448210a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b01553a55ccc3121476a5fdfed5532047ecd1e46545aab873da4a7d4718bd96f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b961d759f8b153e4c876e6f1b5eff499fc72a3c828610fc980cba592cac76015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5aaf34113a5a5c1948b672848f8bbcbc0ef69e6a3efe03202c9210016a0c6980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aaf34113a5a5c1948b672848f8bbcbc0ef69e6a3efe03202c9210016a0c6980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.565860 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.582724 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.602827 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.613989 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.614052 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.614069 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.614490 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.614528 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:48Z","lastTransitionTime":"2025-11-25T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.628931 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.647004 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60a7f9a29a1e574b4eb0fc72705f8c96b6f613f8a7c164605070077212854cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.717830 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.717888 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.717898 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.717915 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.717927 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:48Z","lastTransitionTime":"2025-11-25T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.800719 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.800783 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.800830 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.800920 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:48 crc kubenswrapper[4702]: E1125 13:47:48.801142 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:47:48 crc kubenswrapper[4702]: E1125 13:47:48.801336 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:47:48 crc kubenswrapper[4702]: E1125 13:47:48.801552 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:47:48 crc kubenswrapper[4702]: E1125 13:47:48.801672 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fm2cr" podUID="88665994-5305-4de7-8b69-e38b38d5f87c" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.821729 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.821802 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.821821 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.821854 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.821876 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:48Z","lastTransitionTime":"2025-11-25T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.924428 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.924485 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.924507 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.924529 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:48 crc kubenswrapper[4702]: I1125 13:47:48.924542 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:48Z","lastTransitionTime":"2025-11-25T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.027959 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.028041 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.028059 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.028086 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.028102 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:49Z","lastTransitionTime":"2025-11-25T13:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.130845 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.130902 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.130913 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.130929 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.130941 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:49Z","lastTransitionTime":"2025-11-25T13:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.232809 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.232856 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.232870 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.232887 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.232899 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:49Z","lastTransitionTime":"2025-11-25T13:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.335637 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.335726 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.335744 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.335773 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.335795 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:49Z","lastTransitionTime":"2025-11-25T13:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.439298 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.439355 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.439365 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.439380 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.439390 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:49Z","lastTransitionTime":"2025-11-25T13:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.541992 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.542054 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.542071 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.542096 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.542109 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:49Z","lastTransitionTime":"2025-11-25T13:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.644971 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.645022 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.645035 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.645054 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.645100 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:49Z","lastTransitionTime":"2025-11-25T13:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.747386 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.747447 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.747458 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.747478 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.747490 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:49Z","lastTransitionTime":"2025-11-25T13:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.849912 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.849971 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.849986 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.850004 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.850017 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:49Z","lastTransitionTime":"2025-11-25T13:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.952397 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.952443 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.952452 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.952466 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:49 crc kubenswrapper[4702]: I1125 13:47:49.952476 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:49Z","lastTransitionTime":"2025-11-25T13:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.055325 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.055383 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.055396 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.055414 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.055427 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:50Z","lastTransitionTime":"2025-11-25T13:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.157966 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.158039 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.158055 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.158083 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.158103 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:50Z","lastTransitionTime":"2025-11-25T13:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.261054 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.261139 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.261158 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.261185 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.261206 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:50Z","lastTransitionTime":"2025-11-25T13:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.363691 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.363783 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.363802 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.363831 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.363850 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:50Z","lastTransitionTime":"2025-11-25T13:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.467865 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.467923 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.467936 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.467957 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.467973 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:50Z","lastTransitionTime":"2025-11-25T13:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.572506 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.572641 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.572662 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.572693 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.572712 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:50Z","lastTransitionTime":"2025-11-25T13:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.676336 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.676405 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.676418 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.676438 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.676453 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:50Z","lastTransitionTime":"2025-11-25T13:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.779902 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.779946 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.779954 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.779971 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.779981 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:50Z","lastTransitionTime":"2025-11-25T13:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.800715 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.800748 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.800895 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:50 crc kubenswrapper[4702]: E1125 13:47:50.800931 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fm2cr" podUID="88665994-5305-4de7-8b69-e38b38d5f87c" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.800955 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:47:50 crc kubenswrapper[4702]: E1125 13:47:50.801148 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:47:50 crc kubenswrapper[4702]: E1125 13:47:50.801309 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:47:50 crc kubenswrapper[4702]: E1125 13:47:50.801474 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.883651 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.883737 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.883764 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.883799 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.883819 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:50Z","lastTransitionTime":"2025-11-25T13:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.986939 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.986984 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.986993 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.987010 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:50 crc kubenswrapper[4702]: I1125 13:47:50.987019 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:50Z","lastTransitionTime":"2025-11-25T13:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.090676 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.090741 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.090756 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.090781 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.090799 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:51Z","lastTransitionTime":"2025-11-25T13:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.200693 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.200763 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.200776 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.200796 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.200808 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:51Z","lastTransitionTime":"2025-11-25T13:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.303214 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.303319 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.303338 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.303364 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.303382 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:51Z","lastTransitionTime":"2025-11-25T13:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.407954 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.408004 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.408018 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.408041 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.408056 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:51Z","lastTransitionTime":"2025-11-25T13:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.511361 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.511421 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.511435 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.511460 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.511475 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:51Z","lastTransitionTime":"2025-11-25T13:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.614333 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.614384 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.614397 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.614413 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.614425 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:51Z","lastTransitionTime":"2025-11-25T13:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.718081 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.718173 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.718194 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.718278 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.718298 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:51Z","lastTransitionTime":"2025-11-25T13:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.817135 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88cd2da0-a3ca-40cf-9afb-28522fe24930\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff40c37e9756ca143489aafc21e756e213d7204f041b8b807eef932de448210a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b01553a55ccc3121476a5fdfed5532047ecd1e46545aab873da4a7d4718bd96f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b961d759f8b153e4c876e6f1b5eff499fc72a3c828610fc980cba592cac76015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5aaf34113a5a5c1948b672848f8bbcbc0ef69e6a3efe03202c9210016a0c6980\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aaf34113a5a5c1948b672848f8bbcbc0ef69e6a3efe03202c9210016a0c6980\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:51Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.831062 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.831456 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.831477 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.831522 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.831535 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:51Z","lastTransitionTime":"2025-11-25T13:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.833691 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:51Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.849393 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:51Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.865534 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c491818-b0bb-4d82-a031-96e2dfac8c27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a375acb1bd145421cb3b8288e6af2e56f303008d83eadf0f602253fcde552a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vgszd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qqr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:51Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.891908 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a8fa25e9591d484f8dcb56974242924d4f31c7aef86b4376092842e57e1653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d1bfac94a5fe6b382a9a3783250d6758603d00db9ce12ea44a3b54e6018103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c91a144aef48f19006041105594111c382420f9e6556a64580f6952990ef4db4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2358ac7b25088c8d1b1cc5648a1fcbfd15211826d4c47f2c98839b9924138b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9649c45bc796d60538bc1edc8f1239c7c5445de7b2f3dbf57dbe7df6ecab7d6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5dfe629efca9398d6dcc541b5315c5b6d8e036cba1c94064883785b2cc7a1a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e042cb58071abb8cf4a711485af66c8fa3259aca650ac4a09cb7d66e893c1cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e042cb58071abb8cf4a711485af66c8fa3259aca650ac4a09cb7d66e893c1cfc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T13:47:28Z\\\",\\\"message\\\":\\\"Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 13:47:28.615062 6389 services_controller.go:445] Built service openshift-image-registry/image-registry LB template configs for network=default: []services.lbConfig(nil)\\\\nF1125 13:47:28.615095 6389 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0x1fcbc00 0x1fcbba0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:28Z is after 2025-08-24T17:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p9c5v_openshift-ovn-kubernetes(3c8a3234-01d3-48a9-a5b4-acb46b4218dc)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8700d12c2e236e704fc60136a51489d1d6890a83621480c23a6168d4a0f18f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p9c5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:51Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.906454 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fm2cr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88665994-5305-4de7-8b69-e38b38d5f87c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q8kvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fm2cr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:51Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.920203 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0f51e5a-21b4-4b12-884d-823946038ffd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90db72f00d2ebd6a338e6ec5396349087083c2ad55dd120e712b1e4916b7c5e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e08782f884652e08e363cd5cad0db72b87dedb7eaefa8db537a43ebdfc778325\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e08782f884652e08e363cd5cad0db72b87dedb7eaefa8db537a43ebdfc778325\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:51Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.935070 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.935104 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.935116 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.935135 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.935150 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:51Z","lastTransitionTime":"2025-11-25T13:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.939733 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:51Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.955447 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84e55012afa50a6b8a84745fa4ba0a67e6d8a0f561e94e6a7069cc77ac9359f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:51Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.973850 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://864141de1584fdae9a15214969d215f7479125842181b98e9a589939a1f6fd6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adda7ae248bf46a36e7bda7bde5e11cc6dde217c3e2b295591cea590960855fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:51Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:51 crc kubenswrapper[4702]: I1125 13:47:51.995998 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b142b34b-79ab-4995-9578-d8fdf04e1fba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60a7f9a29a1e574b4eb0fc72705f8c96b6f613f8a7c164605070077212854cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee2658d1a7521e5e617923dba37ba2fb93daa22c2d5ad0ad0264c3b4372711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15d6e5308e3b2939022706ab7c8c66eb996b14a79b7f4e5bb43b89c2cef43ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e17204d9420fb62489c3cf425195c6c5dc50d15723bfa1257bc10a403ffbd5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06cfc0c6778792be5558a9c090f4e08c88ed62b4463d1df09fa8063d0d6f57f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b81f45e9f546f944ab976a2dfe533a1535f85fce00b0709e0490d51af7489ad9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6279085cd312c51925d8778e38e1ae556a44e9bad88a24cc6c18a2241375a662\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:47:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:47:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmrxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wcs7x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:51Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.012756 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f0989940-684e-4f56-af09-3e3d238e16d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://776d82413b70e51c0e8149ac70c57542d11e6a7088054bd6522b447deff8462a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ed6875eaa6db535132f7074ee248583f06ca4bc1f01f7094ec5c986c0eb658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dec9fd3752ee0889181099112317055d318dd29d996d844a5a40cf36c0564c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:52Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.029536 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77def600-8d7d-496f-9f98-d67c0f30305a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T13:46:50Z\\\",\\\"message\\\":\\\"calhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764078394\\\\\\\\\\\\\\\" (2025-11-25 13:46:34 +0000 UTC to 2025-12-25 13:46:35 +0000 UTC (now=2025-11-25 13:46:50.522988783 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523161 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523210 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764078395\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764078395\\\\\\\\\\\\\\\" (2025-11-25 12:46:35 +0000 UTC to 2026-11-25 12:46:35 +0000 UTC (now=2025-11-25 13:46:50.523185508 +0000 UTC))\\\\\\\"\\\\nI1125 13:46:50.523252 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 13:46:50.523278 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 13:46:50.523302 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-181896662/tls.crt::/tmp/serving-cert-181896662/tls.key\\\\\\\"\\\\nI1125 13:46:50.523302 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1125 13:46:50.523393 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523422 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 13:46:50.523440 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1125 13:46:50.523483 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 13:46:50.523508 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1125 13:46:50.523593 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:52Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.037123 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.037155 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.037165 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.037180 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.037193 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:52Z","lastTransitionTime":"2025-11-25T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.042069 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d16085ef8ba07bccefb9f55835dc0d57b07bbc91a89a5bb8c37538ed482b9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:52Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.051742 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5mp5w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83c1c667-d798-4bdc-83d0-25c77a44fc43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c90f0118f854de5973a3e5ad0f983539a9e66bf6195700ab5335a45dcc3720fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnsck\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5mp5w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:52Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.067834 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gd2xs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e735a90-cf24-47de-b0e6-57d27a971b21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10ef8fba2e92332ed66f05f36bc57004f1d12c53689621106a3606f71d8aee9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T13:47:46Z\\\",\\\"message\\\":\\\"2025-11-25T13:47:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_48eac4c0-a560-4b00-86f2-65d9e25b2f5b\\\\n2025-11-25T13:47:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_48eac4c0-a560-4b00-86f2-65d9e25b2f5b to /host/opt/cni/bin/\\\\n2025-11-25T13:47:01Z [verbose] multus-daemon started\\\\n2025-11-25T13:47:01Z [verbose] Readiness Indicator file check\\\\n2025-11-25T13:47:46Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:58Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6sj98\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gd2xs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:52Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.080684 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bhpwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e607da52-3cea-4ebe-b9e4-4eef40d6e714\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d2efcd45f14690e23b56c19e3282342933f83acd4940f89dd4046b61e4a608b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6jwz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bhpwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:52Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.092153 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675f4b11-d943-4acc-afde-6f4f45585551\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35b52152fd6304d8c2c961db7d7b144f4c17d8ee855f67e25ca711a89c60a85d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f010c83c609ad0554ed60899b53504cd7d859528a4947f45356466d15930c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:47:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z6jjm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:47:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-djssv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:52Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.108872 4702 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa6bb46d-9671-44bd-9f1b-8de9e777bc6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T13:46:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://560b3148a88cf519409f54ef1ba3c5ca29bd137150cd854918d0ec39392d1926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15685943c1477944bfb3c22c12358e0f245be68503e6066b5b94ad82780391d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f331ecc144ad33484050b9dc5d17b92ddd558da44ddefd62b11e5c2fe4c8dc7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://834026f07cbf06f9ce28a3ba398dea36d34af50962282b7d04c1d3229112908d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69f05dbc702b71b2e4a07a158000c09fa0b896e65acc337fc0225b3eca7d376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T13:46:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0972bcd308f5ae8eb5420507ba34e71d13c1bd99a781a044421be4d233107512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7452d1cc4032fe43b7d9dc0bb732b09e5d26e3df22e8245fec515b9428af158\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49246d08764da171475b9962154e8ba4bc20995c1c13e74441302ebd1a841f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T13:46:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T13:46:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T13:46:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T13:47:52Z is after 2025-08-24T17:21:41Z" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.142938 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.142980 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.143001 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.143023 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.143043 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:52Z","lastTransitionTime":"2025-11-25T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.246541 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.246870 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.246882 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.246898 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.246909 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:52Z","lastTransitionTime":"2025-11-25T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.349070 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.349104 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.349114 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.349127 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.349136 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:52Z","lastTransitionTime":"2025-11-25T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.452644 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.452722 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.452749 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.452778 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.452799 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:52Z","lastTransitionTime":"2025-11-25T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.555088 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.555146 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.555161 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.555183 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.555198 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:52Z","lastTransitionTime":"2025-11-25T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.658843 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.658921 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.658942 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.658971 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.658992 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:52Z","lastTransitionTime":"2025-11-25T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.763376 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.763422 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.763431 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.763447 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.763456 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:52Z","lastTransitionTime":"2025-11-25T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.800863 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.800915 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:47:52 crc kubenswrapper[4702]: E1125 13:47:52.801018 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fm2cr" podUID="88665994-5305-4de7-8b69-e38b38d5f87c" Nov 25 13:47:52 crc kubenswrapper[4702]: E1125 13:47:52.801167 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.801226 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:52 crc kubenswrapper[4702]: E1125 13:47:52.801324 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.801396 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:47:52 crc kubenswrapper[4702]: E1125 13:47:52.801478 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.865773 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.865836 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.865850 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.865877 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.865893 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:52Z","lastTransitionTime":"2025-11-25T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.968582 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.968652 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.968665 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.968688 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:52 crc kubenswrapper[4702]: I1125 13:47:52.968702 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:52Z","lastTransitionTime":"2025-11-25T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.071936 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.071999 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.072016 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.072040 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.072057 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:53Z","lastTransitionTime":"2025-11-25T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.175203 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.175283 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.175294 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.175308 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.175316 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:53Z","lastTransitionTime":"2025-11-25T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.278089 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.278145 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.278157 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.278179 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.278198 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:53Z","lastTransitionTime":"2025-11-25T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.380256 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.380289 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.380297 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.380310 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.380320 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:53Z","lastTransitionTime":"2025-11-25T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.482520 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.482597 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.482610 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.482625 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.482636 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:53Z","lastTransitionTime":"2025-11-25T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.585748 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.585788 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.585796 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.585812 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.585823 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:53Z","lastTransitionTime":"2025-11-25T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.688633 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.688991 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.689196 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.689328 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.689469 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:53Z","lastTransitionTime":"2025-11-25T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.792084 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.792151 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.792164 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.792181 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.792193 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:53Z","lastTransitionTime":"2025-11-25T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.894812 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.894856 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.894863 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.894878 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.894886 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:53Z","lastTransitionTime":"2025-11-25T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.997020 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.997049 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.997058 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.997071 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:53 crc kubenswrapper[4702]: I1125 13:47:53.997080 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:53Z","lastTransitionTime":"2025-11-25T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.100576 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.100668 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.100699 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.100748 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.100773 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:54Z","lastTransitionTime":"2025-11-25T13:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.204812 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.204885 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.204911 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.204944 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.204969 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:54Z","lastTransitionTime":"2025-11-25T13:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.307079 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.307135 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.307147 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.307166 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.307179 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:54Z","lastTransitionTime":"2025-11-25T13:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.410819 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.410887 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.410909 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.410940 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.410964 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:54Z","lastTransitionTime":"2025-11-25T13:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.514149 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.515114 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.515337 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.515522 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.515679 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:54Z","lastTransitionTime":"2025-11-25T13:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.621732 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.621783 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.621804 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.621835 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.621858 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:54Z","lastTransitionTime":"2025-11-25T13:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.640268 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.640402 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.640425 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:54 crc kubenswrapper[4702]: E1125 13:47:54.640585 4702 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 13:47:54 crc kubenswrapper[4702]: E1125 13:47:54.640641 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 13:48:58.640628795 +0000 UTC m=+147.803279245 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 13:47:54 crc kubenswrapper[4702]: E1125 13:47:54.640645 4702 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 13:47:54 crc kubenswrapper[4702]: E1125 13:47:54.640773 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 13:48:58.640735037 +0000 UTC m=+147.803385537 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 13:47:54 crc kubenswrapper[4702]: E1125 13:47:54.640944 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:58.640918762 +0000 UTC m=+147.803569282 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.725343 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.725424 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.725443 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.725473 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.725500 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:54Z","lastTransitionTime":"2025-11-25T13:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.741478 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.741563 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:47:54 crc kubenswrapper[4702]: E1125 13:47:54.741757 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 13:47:54 crc kubenswrapper[4702]: E1125 13:47:54.741818 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 13:47:54 crc kubenswrapper[4702]: E1125 13:47:54.741823 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 13:47:54 crc kubenswrapper[4702]: E1125 13:47:54.741845 4702 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 13:47:54 crc kubenswrapper[4702]: E1125 13:47:54.741859 4702 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 13:47:54 crc kubenswrapper[4702]: E1125 13:47:54.741868 4702 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 13:47:54 crc kubenswrapper[4702]: E1125 13:47:54.741934 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 13:48:58.741911666 +0000 UTC m=+147.904562156 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 13:47:54 crc kubenswrapper[4702]: E1125 13:47:54.741965 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 13:48:58.741951137 +0000 UTC m=+147.904601627 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.800899 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.800968 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.801053 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.800908 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:47:54 crc kubenswrapper[4702]: E1125 13:47:54.801082 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:47:54 crc kubenswrapper[4702]: E1125 13:47:54.801170 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:47:54 crc kubenswrapper[4702]: E1125 13:47:54.801344 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fm2cr" podUID="88665994-5305-4de7-8b69-e38b38d5f87c" Nov 25 13:47:54 crc kubenswrapper[4702]: E1125 13:47:54.801704 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.828742 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.828789 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.828801 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.828819 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.828835 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:54Z","lastTransitionTime":"2025-11-25T13:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.932131 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.932171 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.932180 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.932194 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:54 crc kubenswrapper[4702]: I1125 13:47:54.932206 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:54Z","lastTransitionTime":"2025-11-25T13:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.034899 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.034935 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.034953 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.034973 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.034985 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:55Z","lastTransitionTime":"2025-11-25T13:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.138104 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.138167 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.138185 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.138213 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.138268 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:55Z","lastTransitionTime":"2025-11-25T13:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.240800 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.240870 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.240887 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.240912 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.240933 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:55Z","lastTransitionTime":"2025-11-25T13:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.344152 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.344205 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.344216 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.344279 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.344293 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:55Z","lastTransitionTime":"2025-11-25T13:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.447077 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.447125 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.447140 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.447156 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.447165 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:55Z","lastTransitionTime":"2025-11-25T13:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.549780 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.549839 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.549849 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.549876 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.549888 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:55Z","lastTransitionTime":"2025-11-25T13:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.653344 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.653383 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.653398 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.653419 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.653435 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:55Z","lastTransitionTime":"2025-11-25T13:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.757213 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.757365 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.757395 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.757424 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.757443 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:55Z","lastTransitionTime":"2025-11-25T13:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.860525 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.860578 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.860592 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.860609 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.860622 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:55Z","lastTransitionTime":"2025-11-25T13:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.963880 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.963932 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.963941 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.963957 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:55 crc kubenswrapper[4702]: I1125 13:47:55.963968 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:55Z","lastTransitionTime":"2025-11-25T13:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.066217 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.066276 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.066285 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.066299 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.066308 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:56Z","lastTransitionTime":"2025-11-25T13:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.168016 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.168053 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.168069 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.168089 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.168100 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:56Z","lastTransitionTime":"2025-11-25T13:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.270756 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.270796 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.270805 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.270817 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.270826 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:56Z","lastTransitionTime":"2025-11-25T13:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.372557 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.372596 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.372628 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.372644 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.372654 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:56Z","lastTransitionTime":"2025-11-25T13:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.474489 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.474581 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.474597 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.474618 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.474634 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:56Z","lastTransitionTime":"2025-11-25T13:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.576705 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.576807 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.576823 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.576843 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.576855 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:56Z","lastTransitionTime":"2025-11-25T13:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.681556 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.681606 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.681634 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.681654 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.681668 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:56Z","lastTransitionTime":"2025-11-25T13:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.784082 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.784124 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.784135 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.784154 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.784169 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:56Z","lastTransitionTime":"2025-11-25T13:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.800699 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.800712 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.801144 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:56 crc kubenswrapper[4702]: E1125 13:47:56.801073 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fm2cr" podUID="88665994-5305-4de7-8b69-e38b38d5f87c" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.801278 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:47:56 crc kubenswrapper[4702]: E1125 13:47:56.801367 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:47:56 crc kubenswrapper[4702]: E1125 13:47:56.801418 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:47:56 crc kubenswrapper[4702]: E1125 13:47:56.801487 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.887700 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.887774 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.887791 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.887820 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.887840 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:56Z","lastTransitionTime":"2025-11-25T13:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.989965 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.990000 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.990008 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.990022 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:56 crc kubenswrapper[4702]: I1125 13:47:56.990031 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:56Z","lastTransitionTime":"2025-11-25T13:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.092429 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.092524 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.092537 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.092559 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.092574 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:57Z","lastTransitionTime":"2025-11-25T13:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.195796 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.195847 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.195859 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.195877 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.195896 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:57Z","lastTransitionTime":"2025-11-25T13:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.299745 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.299797 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.299810 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.299834 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.299848 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:57Z","lastTransitionTime":"2025-11-25T13:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.402757 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.402796 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.402804 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.402817 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.402828 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:57Z","lastTransitionTime":"2025-11-25T13:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.505969 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.506071 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.506114 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.506152 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.506179 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:57Z","lastTransitionTime":"2025-11-25T13:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.609991 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.610075 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.610102 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.610137 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.610165 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:57Z","lastTransitionTime":"2025-11-25T13:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.714015 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.714082 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.714108 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.714143 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.714168 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:57Z","lastTransitionTime":"2025-11-25T13:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.816898 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.816971 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.816991 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.817016 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.817039 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:57Z","lastTransitionTime":"2025-11-25T13:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.920500 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.920576 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.920593 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.920622 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:57 crc kubenswrapper[4702]: I1125 13:47:57.920641 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:57Z","lastTransitionTime":"2025-11-25T13:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.023672 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.023757 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.023784 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.023820 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.023846 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:58Z","lastTransitionTime":"2025-11-25T13:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.126706 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.126754 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.126764 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.126811 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.126823 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:58Z","lastTransitionTime":"2025-11-25T13:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.229877 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.229926 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.229938 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.229956 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.229968 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:58Z","lastTransitionTime":"2025-11-25T13:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.332493 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.332568 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.332590 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.332620 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.332640 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:58Z","lastTransitionTime":"2025-11-25T13:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.405599 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.405650 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.405662 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.405681 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.405693 4702 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T13:47:58Z","lastTransitionTime":"2025-11-25T13:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.455348 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-4h7d6"] Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.455767 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4h7d6" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.458630 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.458638 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.461102 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.461227 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.481832 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/d701b75e-9e65-43e4-a47a-00606a3b45f4-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-4h7d6\" (UID: \"d701b75e-9e65-43e4-a47a-00606a3b45f4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4h7d6" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.481882 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d701b75e-9e65-43e4-a47a-00606a3b45f4-service-ca\") pod \"cluster-version-operator-5c965bbfc6-4h7d6\" (UID: \"d701b75e-9e65-43e4-a47a-00606a3b45f4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4h7d6" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.481955 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d701b75e-9e65-43e4-a47a-00606a3b45f4-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-4h7d6\" (UID: \"d701b75e-9e65-43e4-a47a-00606a3b45f4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4h7d6" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.481980 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/d701b75e-9e65-43e4-a47a-00606a3b45f4-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-4h7d6\" (UID: \"d701b75e-9e65-43e4-a47a-00606a3b45f4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4h7d6" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.482097 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d701b75e-9e65-43e4-a47a-00606a3b45f4-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-4h7d6\" (UID: \"d701b75e-9e65-43e4-a47a-00606a3b45f4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4h7d6" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.525163 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podStartSLOduration=63.525142739 podStartE2EDuration="1m3.525142739s" podCreationTimestamp="2025-11-25 13:46:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:47:58.524686288 +0000 UTC m=+87.687336738" watchObservedRunningTime="2025-11-25 13:47:58.525142739 +0000 UTC m=+87.687793189" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.583729 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/d701b75e-9e65-43e4-a47a-00606a3b45f4-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-4h7d6\" (UID: \"d701b75e-9e65-43e4-a47a-00606a3b45f4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4h7d6" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.584182 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d701b75e-9e65-43e4-a47a-00606a3b45f4-service-ca\") pod \"cluster-version-operator-5c965bbfc6-4h7d6\" (UID: \"d701b75e-9e65-43e4-a47a-00606a3b45f4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4h7d6" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.584380 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d701b75e-9e65-43e4-a47a-00606a3b45f4-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-4h7d6\" (UID: \"d701b75e-9e65-43e4-a47a-00606a3b45f4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4h7d6" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.584485 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/d701b75e-9e65-43e4-a47a-00606a3b45f4-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-4h7d6\" (UID: \"d701b75e-9e65-43e4-a47a-00606a3b45f4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4h7d6" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.584571 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/d701b75e-9e65-43e4-a47a-00606a3b45f4-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-4h7d6\" (UID: \"d701b75e-9e65-43e4-a47a-00606a3b45f4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4h7d6" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.583883 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/d701b75e-9e65-43e4-a47a-00606a3b45f4-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-4h7d6\" (UID: \"d701b75e-9e65-43e4-a47a-00606a3b45f4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4h7d6" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.584593 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d701b75e-9e65-43e4-a47a-00606a3b45f4-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-4h7d6\" (UID: \"d701b75e-9e65-43e4-a47a-00606a3b45f4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4h7d6" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.585181 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d701b75e-9e65-43e4-a47a-00606a3b45f4-service-ca\") pod \"cluster-version-operator-5c965bbfc6-4h7d6\" (UID: \"d701b75e-9e65-43e4-a47a-00606a3b45f4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4h7d6" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.594048 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=11.594024284 podStartE2EDuration="11.594024284s" podCreationTimestamp="2025-11-25 13:47:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:47:58.581372763 +0000 UTC m=+87.744023233" watchObservedRunningTime="2025-11-25 13:47:58.594024284 +0000 UTC m=+87.756674734" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.597124 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d701b75e-9e65-43e4-a47a-00606a3b45f4-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-4h7d6\" (UID: \"d701b75e-9e65-43e4-a47a-00606a3b45f4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4h7d6" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.601499 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d701b75e-9e65-43e4-a47a-00606a3b45f4-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-4h7d6\" (UID: \"d701b75e-9e65-43e4-a47a-00606a3b45f4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4h7d6" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.610593 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=31.610562831 podStartE2EDuration="31.610562831s" podCreationTimestamp="2025-11-25 13:47:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:47:58.59510181 +0000 UTC m=+87.757752270" watchObservedRunningTime="2025-11-25 13:47:58.610562831 +0000 UTC m=+87.773213291" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.705071 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=62.705039886 podStartE2EDuration="1m2.705039886s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:47:58.704529213 +0000 UTC m=+87.867179663" watchObservedRunningTime="2025-11-25 13:47:58.705039886 +0000 UTC m=+87.867690416" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.720714 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-wcs7x" podStartSLOduration=62.720690001 podStartE2EDuration="1m2.720690001s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:47:58.719776708 +0000 UTC m=+87.882427158" watchObservedRunningTime="2025-11-25 13:47:58.720690001 +0000 UTC m=+87.883340461" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.773400 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4h7d6" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.784426 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-5mp5w" podStartSLOduration=63.784396358 podStartE2EDuration="1m3.784396358s" podCreationTimestamp="2025-11-25 13:46:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:47:58.768078477 +0000 UTC m=+87.930728917" watchObservedRunningTime="2025-11-25 13:47:58.784396358 +0000 UTC m=+87.947046818" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.797567 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-gd2xs" podStartSLOduration=62.797540682 podStartE2EDuration="1m2.797540682s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:47:58.784596183 +0000 UTC m=+87.947246643" watchObservedRunningTime="2025-11-25 13:47:58.797540682 +0000 UTC m=+87.960191132" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.800446 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.800454 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.800463 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.800496 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:47:58 crc kubenswrapper[4702]: E1125 13:47:58.801015 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fm2cr" podUID="88665994-5305-4de7-8b69-e38b38d5f87c" Nov 25 13:47:58 crc kubenswrapper[4702]: E1125 13:47:58.801129 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:47:58 crc kubenswrapper[4702]: E1125 13:47:58.801311 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:47:58 crc kubenswrapper[4702]: E1125 13:47:58.801387 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.814813 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-bhpwl" podStartSLOduration=63.814788066 podStartE2EDuration="1m3.814788066s" podCreationTimestamp="2025-11-25 13:46:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:47:58.79788279 +0000 UTC m=+87.960533240" watchObservedRunningTime="2025-11-25 13:47:58.814788066 +0000 UTC m=+87.977438506" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.854504 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-djssv" podStartSLOduration=62.854484942 podStartE2EDuration="1m2.854484942s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:47:58.815678878 +0000 UTC m=+87.978329328" watchObservedRunningTime="2025-11-25 13:47:58.854484942 +0000 UTC m=+88.017135392" Nov 25 13:47:58 crc kubenswrapper[4702]: I1125 13:47:58.855149 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=64.855143689 podStartE2EDuration="1m4.855143689s" podCreationTimestamp="2025-11-25 13:46:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:47:58.854223256 +0000 UTC m=+88.016873706" watchObservedRunningTime="2025-11-25 13:47:58.855143689 +0000 UTC m=+88.017794139" Nov 25 13:47:59 crc kubenswrapper[4702]: I1125 13:47:59.324936 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4h7d6" event={"ID":"d701b75e-9e65-43e4-a47a-00606a3b45f4","Type":"ContainerStarted","Data":"b4203109a58b0ee87bf48ed583d300b3d1121a3615cba25eac05204c0c48ac4c"} Nov 25 13:47:59 crc kubenswrapper[4702]: I1125 13:47:59.801674 4702 scope.go:117] "RemoveContainer" containerID="e042cb58071abb8cf4a711485af66c8fa3259aca650ac4a09cb7d66e893c1cfc" Nov 25 13:48:00 crc kubenswrapper[4702]: I1125 13:48:00.330772 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9c5v_3c8a3234-01d3-48a9-a5b4-acb46b4218dc/ovnkube-controller/2.log" Nov 25 13:48:00 crc kubenswrapper[4702]: I1125 13:48:00.333472 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" event={"ID":"3c8a3234-01d3-48a9-a5b4-acb46b4218dc","Type":"ContainerStarted","Data":"9e437f067d92fb24f70625cdd0e2a221164118fa19cd362de0a7cd5e71e88dd3"} Nov 25 13:48:00 crc kubenswrapper[4702]: I1125 13:48:00.333899 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:48:00 crc kubenswrapper[4702]: I1125 13:48:00.335379 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4h7d6" event={"ID":"d701b75e-9e65-43e4-a47a-00606a3b45f4","Type":"ContainerStarted","Data":"e868d6797a365c831ea8229e93b60b1bcd3f853e91db9d938fab36ca97950261"} Nov 25 13:48:00 crc kubenswrapper[4702]: I1125 13:48:00.378053 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=70.378034397 podStartE2EDuration="1m10.378034397s" podCreationTimestamp="2025-11-25 13:46:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:47:58.876674598 +0000 UTC m=+88.039325058" watchObservedRunningTime="2025-11-25 13:48:00.378034397 +0000 UTC m=+89.540684847" Nov 25 13:48:00 crc kubenswrapper[4702]: I1125 13:48:00.378432 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" podStartSLOduration=64.378426457 podStartE2EDuration="1m4.378426457s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:00.37777395 +0000 UTC m=+89.540424400" watchObservedRunningTime="2025-11-25 13:48:00.378426457 +0000 UTC m=+89.541076907" Nov 25 13:48:00 crc kubenswrapper[4702]: I1125 13:48:00.799017 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4h7d6" podStartSLOduration=64.798995614 podStartE2EDuration="1m4.798995614s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:00.393958589 +0000 UTC m=+89.556609049" watchObservedRunningTime="2025-11-25 13:48:00.798995614 +0000 UTC m=+89.961646064" Nov 25 13:48:00 crc kubenswrapper[4702]: I1125 13:48:00.800188 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-fm2cr"] Nov 25 13:48:00 crc kubenswrapper[4702]: I1125 13:48:00.800368 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:48:00 crc kubenswrapper[4702]: I1125 13:48:00.800417 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:48:00 crc kubenswrapper[4702]: E1125 13:48:00.800463 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:48:00 crc kubenswrapper[4702]: I1125 13:48:00.800480 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:48:00 crc kubenswrapper[4702]: E1125 13:48:00.800574 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:48:00 crc kubenswrapper[4702]: I1125 13:48:00.800379 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:48:00 crc kubenswrapper[4702]: E1125 13:48:00.800717 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fm2cr" podUID="88665994-5305-4de7-8b69-e38b38d5f87c" Nov 25 13:48:00 crc kubenswrapper[4702]: E1125 13:48:00.800776 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:48:02 crc kubenswrapper[4702]: I1125 13:48:02.801004 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:48:02 crc kubenswrapper[4702]: I1125 13:48:02.801089 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:48:02 crc kubenswrapper[4702]: I1125 13:48:02.801004 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:48:02 crc kubenswrapper[4702]: E1125 13:48:02.801187 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 13:48:02 crc kubenswrapper[4702]: E1125 13:48:02.801343 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 13:48:02 crc kubenswrapper[4702]: E1125 13:48:02.801552 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 13:48:02 crc kubenswrapper[4702]: I1125 13:48:02.801606 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:48:02 crc kubenswrapper[4702]: E1125 13:48:02.801674 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fm2cr" podUID="88665994-5305-4de7-8b69-e38b38d5f87c" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.091191 4702 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.091452 4702 kubelet_node_status.go:538] "Fast updating node status as it just became ready" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.130922 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wv6k6"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.131435 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wv6k6" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.134016 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.134616 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.135926 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4zttp"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.136208 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.136333 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4zttp" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.138418 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.138642 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.138767 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.138926 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-c2xs7"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.139338 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wdn7d"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.139534 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-c2xs7" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.139806 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wdn7d" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.140752 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-8j9xk"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.141086 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.141322 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.141107 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-8j9xk" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.141531 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.153556 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.153563 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.153714 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.153955 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.154099 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.154363 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.154381 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.154494 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.154570 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.154681 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.154797 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.155612 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-9bq59"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.156287 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-9bq59" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.156828 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-pvqd2"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.157760 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.158286 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rmtvx"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.159012 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rmtvx" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.162583 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.166314 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.167216 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-mbmx2"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.167921 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mbmx2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.168198 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b2jt8"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.168690 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b2jt8" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.169995 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-dnwh7"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.170280 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-dnwh7" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.173659 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-sxxzq"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.174145 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.174391 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.174557 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.174561 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-dfr9q"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.174650 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.174955 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.175089 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sxxzq" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.175305 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.175346 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.176007 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.176130 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.176450 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.187305 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.187657 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.187807 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.188105 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.188260 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.188641 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.188767 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.190802 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.191072 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.201513 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.201723 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.201747 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.201954 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.202078 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.202146 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.202249 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.202359 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.202430 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.204046 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.204123 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.209813 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.210062 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.210228 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.210412 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.210581 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.211464 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.212327 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.212465 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.212646 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.212759 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.212867 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.213013 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.213078 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9xwkx"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.213513 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-q2bqm"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.213917 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-q2bqm" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.214166 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.214315 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9xwkx" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.214739 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.215173 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.215657 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.215872 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.215994 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.216117 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.216246 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.216361 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.216539 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.216607 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-5tvj9"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.216867 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.217029 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-5tvj9" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.217838 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.218943 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-bjl94"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.219458 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bjl94" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.222605 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.223777 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.223919 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.224047 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-7twx2"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.224660 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-7twx2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.227631 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.227802 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.231258 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.231418 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.231556 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.232401 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.233071 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.233149 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.233311 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.233338 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.233402 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.233479 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.233571 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.233663 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.234165 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.233573 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.235204 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vjbjr"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.235740 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vjbjr" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237047 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26e29bc4-82c6-47c1-8ffa-5e35e4b1c595-config\") pod \"machine-api-operator-5694c8668f-9bq59\" (UID: \"26e29bc4-82c6-47c1-8ffa-5e35e4b1c595\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9bq59" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237074 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5413dfbd-3c8b-42ac-8536-c0331fb7a397-client-ca\") pod \"controller-manager-879f6c89f-rmtvx\" (UID: \"5413dfbd-3c8b-42ac-8536-c0331fb7a397\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rmtvx" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237092 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb22f225-d744-4357-a091-d21d76a31e42-config\") pod \"openshift-apiserver-operator-796bbdcf4f-wdn7d\" (UID: \"bb22f225-d744-4357-a091-d21d76a31e42\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wdn7d" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237110 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtjbt\" (UniqueName: \"kubernetes.io/projected/bb22f225-d744-4357-a091-d21d76a31e42-kube-api-access-qtjbt\") pod \"openshift-apiserver-operator-796bbdcf4f-wdn7d\" (UID: \"bb22f225-d744-4357-a091-d21d76a31e42\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wdn7d" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237126 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d9232dc2-d71f-41b1-864a-82603860af69-proxy-tls\") pod \"machine-config-controller-84d6567774-bjl94\" (UID: \"d9232dc2-d71f-41b1-864a-82603860af69\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bjl94" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237142 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237156 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5413dfbd-3c8b-42ac-8536-c0331fb7a397-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-rmtvx\" (UID: \"5413dfbd-3c8b-42ac-8536-c0331fb7a397\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rmtvx" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237174 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/26e29bc4-82c6-47c1-8ffa-5e35e4b1c595-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-9bq59\" (UID: \"26e29bc4-82c6-47c1-8ffa-5e35e4b1c595\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9bq59" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237193 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5413dfbd-3c8b-42ac-8536-c0331fb7a397-serving-cert\") pod \"controller-manager-879f6c89f-rmtvx\" (UID: \"5413dfbd-3c8b-42ac-8536-c0331fb7a397\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rmtvx" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237252 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hn5lj\" (UniqueName: \"kubernetes.io/projected/7cfd6b45-6486-4b59-9a79-21773426f02b-kube-api-access-hn5lj\") pod \"authentication-operator-69f744f599-dnwh7\" (UID: \"7cfd6b45-6486-4b59-9a79-21773426f02b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dnwh7" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237274 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b619039f-82c2-49ca-9319-75d0883a5a7d-audit-dir\") pod \"apiserver-7bbb656c7d-xsm2s\" (UID: \"b619039f-82c2-49ca-9319-75d0883a5a7d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237325 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/dd97957f-e7df-48c2-b23f-b1101b786766-console-config\") pod \"console-f9d7485db-mbmx2\" (UID: \"dd97957f-e7df-48c2-b23f-b1101b786766\") " pod="openshift-console/console-f9d7485db-mbmx2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237354 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/dd97957f-e7df-48c2-b23f-b1101b786766-oauth-serving-cert\") pod \"console-f9d7485db-mbmx2\" (UID: \"dd97957f-e7df-48c2-b23f-b1101b786766\") " pod="openshift-console/console-f9d7485db-mbmx2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237389 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8b821731-b65f-45df-b510-de915e573be1-trusted-ca-bundle\") pod \"apiserver-76f77b778f-pvqd2\" (UID: \"8b821731-b65f-45df-b510-de915e573be1\") " pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237404 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxpsk\" (UniqueName: \"kubernetes.io/projected/8b821731-b65f-45df-b510-de915e573be1-kube-api-access-fxpsk\") pod \"apiserver-76f77b778f-pvqd2\" (UID: \"8b821731-b65f-45df-b510-de915e573be1\") " pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237424 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237445 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/dd97957f-e7df-48c2-b23f-b1101b786766-console-oauth-config\") pod \"console-f9d7485db-mbmx2\" (UID: \"dd97957f-e7df-48c2-b23f-b1101b786766\") " pod="openshift-console/console-f9d7485db-mbmx2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237470 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/8b821731-b65f-45df-b510-de915e573be1-image-import-ca\") pod \"apiserver-76f77b778f-pvqd2\" (UID: \"8b821731-b65f-45df-b510-de915e573be1\") " pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237490 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237510 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5add632c-b0b2-490d-a0f6-d0193c176134-metrics-tls\") pod \"dns-operator-744455d44c-q2bqm\" (UID: \"5add632c-b0b2-490d-a0f6-d0193c176134\") " pod="openshift-dns-operator/dns-operator-744455d44c-q2bqm" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237532 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/017380dd-ff61-49ce-8603-00daaeef7d5f-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-4zttp\" (UID: \"017380dd-ff61-49ce-8603-00daaeef7d5f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4zttp" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237551 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3ce01e2e-eb84-47e0-a3c9-d4b0044f79a5-auth-proxy-config\") pod \"machine-approver-56656f9798-sxxzq\" (UID: \"3ce01e2e-eb84-47e0-a3c9-d4b0044f79a5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sxxzq" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237567 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237583 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237597 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ce01e2e-eb84-47e0-a3c9-d4b0044f79a5-config\") pod \"machine-approver-56656f9798-sxxzq\" (UID: \"3ce01e2e-eb84-47e0-a3c9-d4b0044f79a5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sxxzq" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237613 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8b821731-b65f-45df-b510-de915e573be1-audit-dir\") pod \"apiserver-76f77b778f-pvqd2\" (UID: \"8b821731-b65f-45df-b510-de915e573be1\") " pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237628 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqh98\" (UniqueName: \"kubernetes.io/projected/ce097f6c-2894-4417-b66b-0f30d91f3889-kube-api-access-zqh98\") pod \"downloads-7954f5f757-8j9xk\" (UID: \"ce097f6c-2894-4417-b66b-0f30d91f3889\") " pod="openshift-console/downloads-7954f5f757-8j9xk" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237643 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b619039f-82c2-49ca-9319-75d0883a5a7d-etcd-client\") pod \"apiserver-7bbb656c7d-xsm2s\" (UID: \"b619039f-82c2-49ca-9319-75d0883a5a7d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237660 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8b821731-b65f-45df-b510-de915e573be1-node-pullsecrets\") pod \"apiserver-76f77b778f-pvqd2\" (UID: \"8b821731-b65f-45df-b510-de915e573be1\") " pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237677 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/358a66f5-3192-4b20-ba92-37c8a8544209-available-featuregates\") pod \"openshift-config-operator-7777fb866f-c2xs7\" (UID: \"358a66f5-3192-4b20-ba92-37c8a8544209\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-c2xs7" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237694 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/8b821731-b65f-45df-b510-de915e573be1-audit\") pod \"apiserver-76f77b778f-pvqd2\" (UID: \"8b821731-b65f-45df-b510-de915e573be1\") " pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237710 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/25f66de5-fdbd-44dc-911d-976d5c2bbfd3-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-b2jt8\" (UID: \"25f66de5-fdbd-44dc-911d-976d5c2bbfd3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b2jt8" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237728 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8cqf\" (UniqueName: \"kubernetes.io/projected/ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73-kube-api-access-q8cqf\") pod \"route-controller-manager-6576b87f9c-wv6k6\" (UID: \"ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wv6k6" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237761 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/358a66f5-3192-4b20-ba92-37c8a8544209-serving-cert\") pod \"openshift-config-operator-7777fb866f-c2xs7\" (UID: \"358a66f5-3192-4b20-ba92-37c8a8544209\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-c2xs7" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237789 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b619039f-82c2-49ca-9319-75d0883a5a7d-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-xsm2s\" (UID: \"b619039f-82c2-49ca-9319-75d0883a5a7d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237810 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73-client-ca\") pod \"route-controller-manager-6576b87f9c-wv6k6\" (UID: \"ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wv6k6" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237829 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237848 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbs8d\" (UniqueName: \"kubernetes.io/projected/3ce01e2e-eb84-47e0-a3c9-d4b0044f79a5-kube-api-access-nbs8d\") pod \"machine-approver-56656f9798-sxxzq\" (UID: \"3ce01e2e-eb84-47e0-a3c9-d4b0044f79a5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sxxzq" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237863 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b619039f-82c2-49ca-9319-75d0883a5a7d-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-xsm2s\" (UID: \"b619039f-82c2-49ca-9319-75d0883a5a7d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237878 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5413dfbd-3c8b-42ac-8536-c0331fb7a397-config\") pod \"controller-manager-879f6c89f-rmtvx\" (UID: \"5413dfbd-3c8b-42ac-8536-c0331fb7a397\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rmtvx" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237896 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4c1d0e3-c3c8-47f3-ba79-d8aa252b3572-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9xwkx\" (UID: \"e4c1d0e3-c3c8-47f3-ba79-d8aa252b3572\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9xwkx" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237912 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rpfs\" (UniqueName: \"kubernetes.io/projected/d9232dc2-d71f-41b1-864a-82603860af69-kube-api-access-7rpfs\") pod \"machine-config-controller-84d6567774-bjl94\" (UID: \"d9232dc2-d71f-41b1-864a-82603860af69\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bjl94" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237952 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cd67b\" (UniqueName: \"kubernetes.io/projected/ab12e899-20e3-4755-a537-3e54c5761af6-kube-api-access-cd67b\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237968 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ccsm\" (UniqueName: \"kubernetes.io/projected/5413dfbd-3c8b-42ac-8536-c0331fb7a397-kube-api-access-7ccsm\") pod \"controller-manager-879f6c89f-rmtvx\" (UID: \"5413dfbd-3c8b-42ac-8536-c0331fb7a397\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rmtvx" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.237987 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b619039f-82c2-49ca-9319-75d0883a5a7d-audit-policies\") pod \"apiserver-7bbb656c7d-xsm2s\" (UID: \"b619039f-82c2-49ca-9319-75d0883a5a7d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.238010 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e4c1d0e3-c3c8-47f3-ba79-d8aa252b3572-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9xwkx\" (UID: \"e4c1d0e3-c3c8-47f3-ba79-d8aa252b3572\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9xwkx" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.238035 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb22f225-d744-4357-a091-d21d76a31e42-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-wdn7d\" (UID: \"bb22f225-d744-4357-a091-d21d76a31e42\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wdn7d" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.238052 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.238067 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dd97957f-e7df-48c2-b23f-b1101b786766-trusted-ca-bundle\") pod \"console-f9d7485db-mbmx2\" (UID: \"dd97957f-e7df-48c2-b23f-b1101b786766\") " pod="openshift-console/console-f9d7485db-mbmx2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.238084 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9l8v\" (UniqueName: \"kubernetes.io/projected/358a66f5-3192-4b20-ba92-37c8a8544209-kube-api-access-r9l8v\") pod \"openshift-config-operator-7777fb866f-c2xs7\" (UID: \"358a66f5-3192-4b20-ba92-37c8a8544209\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-c2xs7" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.238100 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b619039f-82c2-49ca-9319-75d0883a5a7d-serving-cert\") pod \"apiserver-7bbb656c7d-xsm2s\" (UID: \"b619039f-82c2-49ca-9319-75d0883a5a7d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.238120 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/25f66de5-fdbd-44dc-911d-976d5c2bbfd3-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-b2jt8\" (UID: \"25f66de5-fdbd-44dc-911d-976d5c2bbfd3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b2jt8" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.238135 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7cfd6b45-6486-4b59-9a79-21773426f02b-serving-cert\") pod \"authentication-operator-69f744f599-dnwh7\" (UID: \"7cfd6b45-6486-4b59-9a79-21773426f02b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dnwh7" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.238150 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a0fafb6-8df4-457c-9651-1d40cc9742b3-config\") pod \"service-ca-operator-777779d784-5tvj9\" (UID: \"7a0fafb6-8df4-457c-9651-1d40cc9742b3\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5tvj9" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.238167 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g96tq\" (UniqueName: \"kubernetes.io/projected/7a0fafb6-8df4-457c-9651-1d40cc9742b3-kube-api-access-g96tq\") pod \"service-ca-operator-777779d784-5tvj9\" (UID: \"7a0fafb6-8df4-457c-9651-1d40cc9742b3\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5tvj9" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.238180 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsgkj\" (UniqueName: \"kubernetes.io/projected/dd97957f-e7df-48c2-b23f-b1101b786766-kube-api-access-bsgkj\") pod \"console-f9d7485db-mbmx2\" (UID: \"dd97957f-e7df-48c2-b23f-b1101b786766\") " pod="openshift-console/console-f9d7485db-mbmx2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.238194 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ab12e899-20e3-4755-a537-3e54c5761af6-audit-dir\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.238208 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cfd6b45-6486-4b59-9a79-21773426f02b-config\") pod \"authentication-operator-69f744f599-dnwh7\" (UID: \"7cfd6b45-6486-4b59-9a79-21773426f02b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dnwh7" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.238223 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73-config\") pod \"route-controller-manager-6576b87f9c-wv6k6\" (UID: \"ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wv6k6" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.248830 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b619039f-82c2-49ca-9319-75d0883a5a7d-encryption-config\") pod \"apiserver-7bbb656c7d-xsm2s\" (UID: \"b619039f-82c2-49ca-9319-75d0883a5a7d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.248877 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7a0fafb6-8df4-457c-9651-1d40cc9742b3-serving-cert\") pod \"service-ca-operator-777779d784-5tvj9\" (UID: \"7a0fafb6-8df4-457c-9651-1d40cc9742b3\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5tvj9" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.248900 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/26e29bc4-82c6-47c1-8ffa-5e35e4b1c595-images\") pod \"machine-api-operator-5694c8668f-9bq59\" (UID: \"26e29bc4-82c6-47c1-8ffa-5e35e4b1c595\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9bq59" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.248918 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsd8v\" (UniqueName: \"kubernetes.io/projected/26e29bc4-82c6-47c1-8ffa-5e35e4b1c595-kube-api-access-nsd8v\") pod \"machine-api-operator-5694c8668f-9bq59\" (UID: \"26e29bc4-82c6-47c1-8ffa-5e35e4b1c595\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9bq59" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.248938 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b821731-b65f-45df-b510-de915e573be1-serving-cert\") pod \"apiserver-76f77b778f-pvqd2\" (UID: \"8b821731-b65f-45df-b510-de915e573be1\") " pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.248953 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.248967 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/25f66de5-fdbd-44dc-911d-976d5c2bbfd3-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-b2jt8\" (UID: \"25f66de5-fdbd-44dc-911d-976d5c2bbfd3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b2jt8" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.248983 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/dd97957f-e7df-48c2-b23f-b1101b786766-service-ca\") pod \"console-f9d7485db-mbmx2\" (UID: \"dd97957f-e7df-48c2-b23f-b1101b786766\") " pod="openshift-console/console-f9d7485db-mbmx2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.248999 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rjll\" (UniqueName: \"kubernetes.io/projected/017380dd-ff61-49ce-8603-00daaeef7d5f-kube-api-access-8rjll\") pod \"cluster-samples-operator-665b6dd947-4zttp\" (UID: \"017380dd-ff61-49ce-8603-00daaeef7d5f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4zttp" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.249015 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ab12e899-20e3-4755-a537-3e54c5761af6-audit-policies\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.249030 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.249047 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.249062 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gv77\" (UniqueName: \"kubernetes.io/projected/25f66de5-fdbd-44dc-911d-976d5c2bbfd3-kube-api-access-9gv77\") pod \"cluster-image-registry-operator-dc59b4c8b-b2jt8\" (UID: \"25f66de5-fdbd-44dc-911d-976d5c2bbfd3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b2jt8" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.249078 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4c1d0e3-c3c8-47f3-ba79-d8aa252b3572-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9xwkx\" (UID: \"e4c1d0e3-c3c8-47f3-ba79-d8aa252b3572\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9xwkx" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.249094 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkcwj\" (UniqueName: \"kubernetes.io/projected/5add632c-b0b2-490d-a0f6-d0193c176134-kube-api-access-qkcwj\") pod \"dns-operator-744455d44c-q2bqm\" (UID: \"5add632c-b0b2-490d-a0f6-d0193c176134\") " pod="openshift-dns-operator/dns-operator-744455d44c-q2bqm" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.249112 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7cfd6b45-6486-4b59-9a79-21773426f02b-service-ca-bundle\") pod \"authentication-operator-69f744f599-dnwh7\" (UID: \"7cfd6b45-6486-4b59-9a79-21773426f02b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dnwh7" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.249126 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/dd97957f-e7df-48c2-b23f-b1101b786766-console-serving-cert\") pod \"console-f9d7485db-mbmx2\" (UID: \"dd97957f-e7df-48c2-b23f-b1101b786766\") " pod="openshift-console/console-f9d7485db-mbmx2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.249143 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d9232dc2-d71f-41b1-864a-82603860af69-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-bjl94\" (UID: \"d9232dc2-d71f-41b1-864a-82603860af69\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bjl94" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.249162 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73-serving-cert\") pod \"route-controller-manager-6576b87f9c-wv6k6\" (UID: \"ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wv6k6" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.249194 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b821731-b65f-45df-b510-de915e573be1-config\") pod \"apiserver-76f77b778f-pvqd2\" (UID: \"8b821731-b65f-45df-b510-de915e573be1\") " pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.249211 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8b821731-b65f-45df-b510-de915e573be1-encryption-config\") pod \"apiserver-76f77b778f-pvqd2\" (UID: \"8b821731-b65f-45df-b510-de915e573be1\") " pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.249262 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8b821731-b65f-45df-b510-de915e573be1-etcd-client\") pod \"apiserver-76f77b778f-pvqd2\" (UID: \"8b821731-b65f-45df-b510-de915e573be1\") " pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.249284 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8b821731-b65f-45df-b510-de915e573be1-etcd-serving-ca\") pod \"apiserver-76f77b778f-pvqd2\" (UID: \"8b821731-b65f-45df-b510-de915e573be1\") " pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.249302 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jp8wv\" (UniqueName: \"kubernetes.io/projected/b619039f-82c2-49ca-9319-75d0883a5a7d-kube-api-access-jp8wv\") pod \"apiserver-7bbb656c7d-xsm2s\" (UID: \"b619039f-82c2-49ca-9319-75d0883a5a7d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.241836 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.246261 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5vx95"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.249321 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.249569 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/3ce01e2e-eb84-47e0-a3c9-d4b0044f79a5-machine-approver-tls\") pod \"machine-approver-56656f9798-sxxzq\" (UID: \"3ce01e2e-eb84-47e0-a3c9-d4b0044f79a5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sxxzq" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.249592 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7cfd6b45-6486-4b59-9a79-21773426f02b-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-dnwh7\" (UID: \"7cfd6b45-6486-4b59-9a79-21773426f02b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dnwh7" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.241985 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.244402 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.249829 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-twgvd"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.245943 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.249923 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5vx95" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.248326 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.248591 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.258634 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xszrd"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.260315 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mb6kd"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.264088 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-twgvd" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.264721 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xszrd" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.271731 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.274453 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.274751 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.286722 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.287630 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.287630 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.294053 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.296188 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gj88f"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.296396 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mb6kd" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.297454 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-8vr2l"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.297967 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vlz75"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.298401 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-bmb9v"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.298934 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-j6d5l"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.299560 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8vr2l" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.299819 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.299964 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gj88f" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.300284 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-bmb9v" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.302047 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.302258 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-j7l2t"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.302365 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-j6d5l" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.303223 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401305-j5rqf"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.303476 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-j7l2t" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.304557 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t5vtr"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.305136 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-mlxwh"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.305375 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401305-j5rqf" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.305623 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t5vtr" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.306483 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-zqrgn"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.306839 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lkx85"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.306974 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mlxwh" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.307144 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4zsfb"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.307891 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dk89m"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.308203 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-lmq49"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.308439 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-zqrgn" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.308449 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lkx85" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.308640 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wv6k6"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.308668 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4zttp"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.308719 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-lmq49" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.308820 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4zsfb" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.308904 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dk89m" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.309762 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wdn7d"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.310833 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-c2xs7"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.315842 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-9bq59"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.315902 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-mbmx2"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.315917 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-bjl94"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.317513 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rmtvx"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.318310 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-q2bqm"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.319449 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-5cmt4"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.320197 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.320484 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-5cmt4" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.320572 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-jqmtw"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.324973 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-8j9xk"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.325320 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-7twx2"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.325573 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-jqmtw" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.329309 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-dnwh7"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.334463 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.334519 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-pvqd2"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.337053 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.340072 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t5vtr"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.343158 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xszrd"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.349817 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-j7l2t"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350254 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsd8v\" (UniqueName: \"kubernetes.io/projected/26e29bc4-82c6-47c1-8ffa-5e35e4b1c595-kube-api-access-nsd8v\") pod \"machine-api-operator-5694c8668f-9bq59\" (UID: \"26e29bc4-82c6-47c1-8ffa-5e35e4b1c595\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9bq59" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350299 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7a0fafb6-8df4-457c-9651-1d40cc9742b3-serving-cert\") pod \"service-ca-operator-777779d784-5tvj9\" (UID: \"7a0fafb6-8df4-457c-9651-1d40cc9742b3\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5tvj9" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350319 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/26e29bc4-82c6-47c1-8ffa-5e35e4b1c595-images\") pod \"machine-api-operator-5694c8668f-9bq59\" (UID: \"26e29bc4-82c6-47c1-8ffa-5e35e4b1c595\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9bq59" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350335 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b821731-b65f-45df-b510-de915e573be1-serving-cert\") pod \"apiserver-76f77b778f-pvqd2\" (UID: \"8b821731-b65f-45df-b510-de915e573be1\") " pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350353 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350369 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/25f66de5-fdbd-44dc-911d-976d5c2bbfd3-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-b2jt8\" (UID: \"25f66de5-fdbd-44dc-911d-976d5c2bbfd3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b2jt8" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350386 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/dd97957f-e7df-48c2-b23f-b1101b786766-service-ca\") pod \"console-f9d7485db-mbmx2\" (UID: \"dd97957f-e7df-48c2-b23f-b1101b786766\") " pod="openshift-console/console-f9d7485db-mbmx2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350405 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rjll\" (UniqueName: \"kubernetes.io/projected/017380dd-ff61-49ce-8603-00daaeef7d5f-kube-api-access-8rjll\") pod \"cluster-samples-operator-665b6dd947-4zttp\" (UID: \"017380dd-ff61-49ce-8603-00daaeef7d5f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4zttp" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350431 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350449 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gv77\" (UniqueName: \"kubernetes.io/projected/25f66de5-fdbd-44dc-911d-976d5c2bbfd3-kube-api-access-9gv77\") pod \"cluster-image-registry-operator-dc59b4c8b-b2jt8\" (UID: \"25f66de5-fdbd-44dc-911d-976d5c2bbfd3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b2jt8" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350466 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4c1d0e3-c3c8-47f3-ba79-d8aa252b3572-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9xwkx\" (UID: \"e4c1d0e3-c3c8-47f3-ba79-d8aa252b3572\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9xwkx" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350485 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkcwj\" (UniqueName: \"kubernetes.io/projected/5add632c-b0b2-490d-a0f6-d0193c176134-kube-api-access-qkcwj\") pod \"dns-operator-744455d44c-q2bqm\" (UID: \"5add632c-b0b2-490d-a0f6-d0193c176134\") " pod="openshift-dns-operator/dns-operator-744455d44c-q2bqm" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350511 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ab12e899-20e3-4755-a537-3e54c5761af6-audit-policies\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350529 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350544 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d9232dc2-d71f-41b1-864a-82603860af69-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-bjl94\" (UID: \"d9232dc2-d71f-41b1-864a-82603860af69\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bjl94" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350562 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7cfd6b45-6486-4b59-9a79-21773426f02b-service-ca-bundle\") pod \"authentication-operator-69f744f599-dnwh7\" (UID: \"7cfd6b45-6486-4b59-9a79-21773426f02b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dnwh7" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350588 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/dd97957f-e7df-48c2-b23f-b1101b786766-console-serving-cert\") pod \"console-f9d7485db-mbmx2\" (UID: \"dd97957f-e7df-48c2-b23f-b1101b786766\") " pod="openshift-console/console-f9d7485db-mbmx2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350605 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73-serving-cert\") pod \"route-controller-manager-6576b87f9c-wv6k6\" (UID: \"ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wv6k6" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350620 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b821731-b65f-45df-b510-de915e573be1-config\") pod \"apiserver-76f77b778f-pvqd2\" (UID: \"8b821731-b65f-45df-b510-de915e573be1\") " pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350637 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8b821731-b65f-45df-b510-de915e573be1-encryption-config\") pod \"apiserver-76f77b778f-pvqd2\" (UID: \"8b821731-b65f-45df-b510-de915e573be1\") " pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350653 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8b821731-b65f-45df-b510-de915e573be1-etcd-serving-ca\") pod \"apiserver-76f77b778f-pvqd2\" (UID: \"8b821731-b65f-45df-b510-de915e573be1\") " pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350670 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jp8wv\" (UniqueName: \"kubernetes.io/projected/b619039f-82c2-49ca-9319-75d0883a5a7d-kube-api-access-jp8wv\") pod \"apiserver-7bbb656c7d-xsm2s\" (UID: \"b619039f-82c2-49ca-9319-75d0883a5a7d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350687 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8b821731-b65f-45df-b510-de915e573be1-etcd-client\") pod \"apiserver-76f77b778f-pvqd2\" (UID: \"8b821731-b65f-45df-b510-de915e573be1\") " pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350705 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7cfd6b45-6486-4b59-9a79-21773426f02b-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-dnwh7\" (UID: \"7cfd6b45-6486-4b59-9a79-21773426f02b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dnwh7" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350729 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350748 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/3ce01e2e-eb84-47e0-a3c9-d4b0044f79a5-machine-approver-tls\") pod \"machine-approver-56656f9798-sxxzq\" (UID: \"3ce01e2e-eb84-47e0-a3c9-d4b0044f79a5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sxxzq" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350765 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d9232dc2-d71f-41b1-864a-82603860af69-proxy-tls\") pod \"machine-config-controller-84d6567774-bjl94\" (UID: \"d9232dc2-d71f-41b1-864a-82603860af69\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bjl94" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350783 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26e29bc4-82c6-47c1-8ffa-5e35e4b1c595-config\") pod \"machine-api-operator-5694c8668f-9bq59\" (UID: \"26e29bc4-82c6-47c1-8ffa-5e35e4b1c595\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9bq59" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350799 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5413dfbd-3c8b-42ac-8536-c0331fb7a397-client-ca\") pod \"controller-manager-879f6c89f-rmtvx\" (UID: \"5413dfbd-3c8b-42ac-8536-c0331fb7a397\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rmtvx" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350817 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb22f225-d744-4357-a091-d21d76a31e42-config\") pod \"openshift-apiserver-operator-796bbdcf4f-wdn7d\" (UID: \"bb22f225-d744-4357-a091-d21d76a31e42\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wdn7d" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350837 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtjbt\" (UniqueName: \"kubernetes.io/projected/bb22f225-d744-4357-a091-d21d76a31e42-kube-api-access-qtjbt\") pod \"openshift-apiserver-operator-796bbdcf4f-wdn7d\" (UID: \"bb22f225-d744-4357-a091-d21d76a31e42\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wdn7d" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350857 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350876 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5413dfbd-3c8b-42ac-8536-c0331fb7a397-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-rmtvx\" (UID: \"5413dfbd-3c8b-42ac-8536-c0331fb7a397\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rmtvx" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350895 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b619039f-82c2-49ca-9319-75d0883a5a7d-audit-dir\") pod \"apiserver-7bbb656c7d-xsm2s\" (UID: \"b619039f-82c2-49ca-9319-75d0883a5a7d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350913 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/dd97957f-e7df-48c2-b23f-b1101b786766-console-config\") pod \"console-f9d7485db-mbmx2\" (UID: \"dd97957f-e7df-48c2-b23f-b1101b786766\") " pod="openshift-console/console-f9d7485db-mbmx2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350929 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/dd97957f-e7df-48c2-b23f-b1101b786766-oauth-serving-cert\") pod \"console-f9d7485db-mbmx2\" (UID: \"dd97957f-e7df-48c2-b23f-b1101b786766\") " pod="openshift-console/console-f9d7485db-mbmx2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350948 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/26e29bc4-82c6-47c1-8ffa-5e35e4b1c595-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-9bq59\" (UID: \"26e29bc4-82c6-47c1-8ffa-5e35e4b1c595\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9bq59" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350964 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5413dfbd-3c8b-42ac-8536-c0331fb7a397-serving-cert\") pod \"controller-manager-879f6c89f-rmtvx\" (UID: \"5413dfbd-3c8b-42ac-8536-c0331fb7a397\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rmtvx" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.350993 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hn5lj\" (UniqueName: \"kubernetes.io/projected/7cfd6b45-6486-4b59-9a79-21773426f02b-kube-api-access-hn5lj\") pod \"authentication-operator-69f744f599-dnwh7\" (UID: \"7cfd6b45-6486-4b59-9a79-21773426f02b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dnwh7" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.351010 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.351029 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/dd97957f-e7df-48c2-b23f-b1101b786766-console-oauth-config\") pod \"console-f9d7485db-mbmx2\" (UID: \"dd97957f-e7df-48c2-b23f-b1101b786766\") " pod="openshift-console/console-f9d7485db-mbmx2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.351055 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8b821731-b65f-45df-b510-de915e573be1-trusted-ca-bundle\") pod \"apiserver-76f77b778f-pvqd2\" (UID: \"8b821731-b65f-45df-b510-de915e573be1\") " pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.351071 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxpsk\" (UniqueName: \"kubernetes.io/projected/8b821731-b65f-45df-b510-de915e573be1-kube-api-access-fxpsk\") pod \"apiserver-76f77b778f-pvqd2\" (UID: \"8b821731-b65f-45df-b510-de915e573be1\") " pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.351090 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/8b821731-b65f-45df-b510-de915e573be1-image-import-ca\") pod \"apiserver-76f77b778f-pvqd2\" (UID: \"8b821731-b65f-45df-b510-de915e573be1\") " pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.351108 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.351127 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5add632c-b0b2-490d-a0f6-d0193c176134-metrics-tls\") pod \"dns-operator-744455d44c-q2bqm\" (UID: \"5add632c-b0b2-490d-a0f6-d0193c176134\") " pod="openshift-dns-operator/dns-operator-744455d44c-q2bqm" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.351142 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/017380dd-ff61-49ce-8603-00daaeef7d5f-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-4zttp\" (UID: \"017380dd-ff61-49ce-8603-00daaeef7d5f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4zttp" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.351160 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3ce01e2e-eb84-47e0-a3c9-d4b0044f79a5-auth-proxy-config\") pod \"machine-approver-56656f9798-sxxzq\" (UID: \"3ce01e2e-eb84-47e0-a3c9-d4b0044f79a5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sxxzq" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.351179 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ce01e2e-eb84-47e0-a3c9-d4b0044f79a5-config\") pod \"machine-approver-56656f9798-sxxzq\" (UID: \"3ce01e2e-eb84-47e0-a3c9-d4b0044f79a5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sxxzq" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.351197 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.351213 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.351246 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqh98\" (UniqueName: \"kubernetes.io/projected/ce097f6c-2894-4417-b66b-0f30d91f3889-kube-api-access-zqh98\") pod \"downloads-7954f5f757-8j9xk\" (UID: \"ce097f6c-2894-4417-b66b-0f30d91f3889\") " pod="openshift-console/downloads-7954f5f757-8j9xk" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.351264 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b619039f-82c2-49ca-9319-75d0883a5a7d-etcd-client\") pod \"apiserver-7bbb656c7d-xsm2s\" (UID: \"b619039f-82c2-49ca-9319-75d0883a5a7d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.351282 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8b821731-b65f-45df-b510-de915e573be1-audit-dir\") pod \"apiserver-76f77b778f-pvqd2\" (UID: \"8b821731-b65f-45df-b510-de915e573be1\") " pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.351299 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/358a66f5-3192-4b20-ba92-37c8a8544209-available-featuregates\") pod \"openshift-config-operator-7777fb866f-c2xs7\" (UID: \"358a66f5-3192-4b20-ba92-37c8a8544209\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-c2xs7" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.351316 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8b821731-b65f-45df-b510-de915e573be1-node-pullsecrets\") pod \"apiserver-76f77b778f-pvqd2\" (UID: \"8b821731-b65f-45df-b510-de915e573be1\") " pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.351769 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b619039f-82c2-49ca-9319-75d0883a5a7d-audit-dir\") pod \"apiserver-7bbb656c7d-xsm2s\" (UID: \"b619039f-82c2-49ca-9319-75d0883a5a7d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.351865 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/25f66de5-fdbd-44dc-911d-976d5c2bbfd3-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-b2jt8\" (UID: \"25f66de5-fdbd-44dc-911d-976d5c2bbfd3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b2jt8" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.351886 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/8b821731-b65f-45df-b510-de915e573be1-audit\") pod \"apiserver-76f77b778f-pvqd2\" (UID: \"8b821731-b65f-45df-b510-de915e573be1\") " pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.351902 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8cqf\" (UniqueName: \"kubernetes.io/projected/ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73-kube-api-access-q8cqf\") pod \"route-controller-manager-6576b87f9c-wv6k6\" (UID: \"ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wv6k6" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.351918 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/358a66f5-3192-4b20-ba92-37c8a8544209-serving-cert\") pod \"openshift-config-operator-7777fb866f-c2xs7\" (UID: \"358a66f5-3192-4b20-ba92-37c8a8544209\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-c2xs7" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.351934 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b619039f-82c2-49ca-9319-75d0883a5a7d-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-xsm2s\" (UID: \"b619039f-82c2-49ca-9319-75d0883a5a7d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.351950 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73-client-ca\") pod \"route-controller-manager-6576b87f9c-wv6k6\" (UID: \"ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wv6k6" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.351967 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.351985 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbs8d\" (UniqueName: \"kubernetes.io/projected/3ce01e2e-eb84-47e0-a3c9-d4b0044f79a5-kube-api-access-nbs8d\") pod \"machine-approver-56656f9798-sxxzq\" (UID: \"3ce01e2e-eb84-47e0-a3c9-d4b0044f79a5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sxxzq" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.352003 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b619039f-82c2-49ca-9319-75d0883a5a7d-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-xsm2s\" (UID: \"b619039f-82c2-49ca-9319-75d0883a5a7d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.352024 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rpfs\" (UniqueName: \"kubernetes.io/projected/d9232dc2-d71f-41b1-864a-82603860af69-kube-api-access-7rpfs\") pod \"machine-config-controller-84d6567774-bjl94\" (UID: \"d9232dc2-d71f-41b1-864a-82603860af69\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bjl94" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.352043 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5413dfbd-3c8b-42ac-8536-c0331fb7a397-config\") pod \"controller-manager-879f6c89f-rmtvx\" (UID: \"5413dfbd-3c8b-42ac-8536-c0331fb7a397\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rmtvx" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.352059 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4c1d0e3-c3c8-47f3-ba79-d8aa252b3572-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9xwkx\" (UID: \"e4c1d0e3-c3c8-47f3-ba79-d8aa252b3572\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9xwkx" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.352092 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ccsm\" (UniqueName: \"kubernetes.io/projected/5413dfbd-3c8b-42ac-8536-c0331fb7a397-kube-api-access-7ccsm\") pod \"controller-manager-879f6c89f-rmtvx\" (UID: \"5413dfbd-3c8b-42ac-8536-c0331fb7a397\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rmtvx" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.352108 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b619039f-82c2-49ca-9319-75d0883a5a7d-audit-policies\") pod \"apiserver-7bbb656c7d-xsm2s\" (UID: \"b619039f-82c2-49ca-9319-75d0883a5a7d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.352132 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cd67b\" (UniqueName: \"kubernetes.io/projected/ab12e899-20e3-4755-a537-3e54c5761af6-kube-api-access-cd67b\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.352149 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e4c1d0e3-c3c8-47f3-ba79-d8aa252b3572-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9xwkx\" (UID: \"e4c1d0e3-c3c8-47f3-ba79-d8aa252b3572\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9xwkx" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.352166 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb22f225-d744-4357-a091-d21d76a31e42-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-wdn7d\" (UID: \"bb22f225-d744-4357-a091-d21d76a31e42\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wdn7d" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.352187 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.352206 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dd97957f-e7df-48c2-b23f-b1101b786766-trusted-ca-bundle\") pod \"console-f9d7485db-mbmx2\" (UID: \"dd97957f-e7df-48c2-b23f-b1101b786766\") " pod="openshift-console/console-f9d7485db-mbmx2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.352225 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9l8v\" (UniqueName: \"kubernetes.io/projected/358a66f5-3192-4b20-ba92-37c8a8544209-kube-api-access-r9l8v\") pod \"openshift-config-operator-7777fb866f-c2xs7\" (UID: \"358a66f5-3192-4b20-ba92-37c8a8544209\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-c2xs7" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.352263 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b619039f-82c2-49ca-9319-75d0883a5a7d-serving-cert\") pod \"apiserver-7bbb656c7d-xsm2s\" (UID: \"b619039f-82c2-49ca-9319-75d0883a5a7d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.352286 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/25f66de5-fdbd-44dc-911d-976d5c2bbfd3-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-b2jt8\" (UID: \"25f66de5-fdbd-44dc-911d-976d5c2bbfd3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b2jt8" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.352302 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7cfd6b45-6486-4b59-9a79-21773426f02b-serving-cert\") pod \"authentication-operator-69f744f599-dnwh7\" (UID: \"7cfd6b45-6486-4b59-9a79-21773426f02b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dnwh7" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.352318 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g96tq\" (UniqueName: \"kubernetes.io/projected/7a0fafb6-8df4-457c-9651-1d40cc9742b3-kube-api-access-g96tq\") pod \"service-ca-operator-777779d784-5tvj9\" (UID: \"7a0fafb6-8df4-457c-9651-1d40cc9742b3\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5tvj9" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.352340 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsgkj\" (UniqueName: \"kubernetes.io/projected/dd97957f-e7df-48c2-b23f-b1101b786766-kube-api-access-bsgkj\") pod \"console-f9d7485db-mbmx2\" (UID: \"dd97957f-e7df-48c2-b23f-b1101b786766\") " pod="openshift-console/console-f9d7485db-mbmx2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.352364 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a0fafb6-8df4-457c-9651-1d40cc9742b3-config\") pod \"service-ca-operator-777779d784-5tvj9\" (UID: \"7a0fafb6-8df4-457c-9651-1d40cc9742b3\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5tvj9" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.352376 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.352375 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/26e29bc4-82c6-47c1-8ffa-5e35e4b1c595-images\") pod \"machine-api-operator-5694c8668f-9bq59\" (UID: \"26e29bc4-82c6-47c1-8ffa-5e35e4b1c595\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9bq59" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.352428 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ab12e899-20e3-4755-a537-3e54c5761af6-audit-dir\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.353084 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ab12e899-20e3-4755-a537-3e54c5761af6-audit-policies\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.353190 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d9232dc2-d71f-41b1-864a-82603860af69-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-bjl94\" (UID: \"d9232dc2-d71f-41b1-864a-82603860af69\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bjl94" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.352389 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ab12e899-20e3-4755-a537-3e54c5761af6-audit-dir\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.353564 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cfd6b45-6486-4b59-9a79-21773426f02b-config\") pod \"authentication-operator-69f744f599-dnwh7\" (UID: \"7cfd6b45-6486-4b59-9a79-21773426f02b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dnwh7" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.353589 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73-config\") pod \"route-controller-manager-6576b87f9c-wv6k6\" (UID: \"ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wv6k6" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.353611 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b619039f-82c2-49ca-9319-75d0883a5a7d-encryption-config\") pod \"apiserver-7bbb656c7d-xsm2s\" (UID: \"b619039f-82c2-49ca-9319-75d0883a5a7d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.353636 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/dd97957f-e7df-48c2-b23f-b1101b786766-console-config\") pod \"console-f9d7485db-mbmx2\" (UID: \"dd97957f-e7df-48c2-b23f-b1101b786766\") " pod="openshift-console/console-f9d7485db-mbmx2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.353857 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4c1d0e3-c3c8-47f3-ba79-d8aa252b3572-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9xwkx\" (UID: \"e4c1d0e3-c3c8-47f3-ba79-d8aa252b3572\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9xwkx" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.354175 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/dd97957f-e7df-48c2-b23f-b1101b786766-oauth-serving-cert\") pod \"console-f9d7485db-mbmx2\" (UID: \"dd97957f-e7df-48c2-b23f-b1101b786766\") " pod="openshift-console/console-f9d7485db-mbmx2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.354887 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cfd6b45-6486-4b59-9a79-21773426f02b-config\") pod \"authentication-operator-69f744f599-dnwh7\" (UID: \"7cfd6b45-6486-4b59-9a79-21773426f02b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dnwh7" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.355053 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7cfd6b45-6486-4b59-9a79-21773426f02b-service-ca-bundle\") pod \"authentication-operator-69f744f599-dnwh7\" (UID: \"7cfd6b45-6486-4b59-9a79-21773426f02b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dnwh7" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.357066 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/dd97957f-e7df-48c2-b23f-b1101b786766-service-ca\") pod \"console-f9d7485db-mbmx2\" (UID: \"dd97957f-e7df-48c2-b23f-b1101b786766\") " pod="openshift-console/console-f9d7485db-mbmx2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.357213 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b619039f-82c2-49ca-9319-75d0883a5a7d-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-xsm2s\" (UID: \"b619039f-82c2-49ca-9319-75d0883a5a7d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.357301 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.357594 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b619039f-82c2-49ca-9319-75d0883a5a7d-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-xsm2s\" (UID: \"b619039f-82c2-49ca-9319-75d0883a5a7d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.357666 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73-config\") pod \"route-controller-manager-6576b87f9c-wv6k6\" (UID: \"ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wv6k6" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.358146 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb22f225-d744-4357-a091-d21d76a31e42-config\") pod \"openshift-apiserver-operator-796bbdcf4f-wdn7d\" (UID: \"bb22f225-d744-4357-a091-d21d76a31e42\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wdn7d" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.358167 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73-client-ca\") pod \"route-controller-manager-6576b87f9c-wv6k6\" (UID: \"ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wv6k6" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.359094 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.359882 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26e29bc4-82c6-47c1-8ffa-5e35e4b1c595-config\") pod \"machine-api-operator-5694c8668f-9bq59\" (UID: \"26e29bc4-82c6-47c1-8ffa-5e35e4b1c595\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9bq59" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.360602 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5413dfbd-3c8b-42ac-8536-c0331fb7a397-client-ca\") pod \"controller-manager-879f6c89f-rmtvx\" (UID: \"5413dfbd-3c8b-42ac-8536-c0331fb7a397\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rmtvx" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.361318 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/8b821731-b65f-45df-b510-de915e573be1-image-import-ca\") pod \"apiserver-76f77b778f-pvqd2\" (UID: \"8b821731-b65f-45df-b510-de915e573be1\") " pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.361813 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8b821731-b65f-45df-b510-de915e573be1-trusted-ca-bundle\") pod \"apiserver-76f77b778f-pvqd2\" (UID: \"8b821731-b65f-45df-b510-de915e573be1\") " pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.362046 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5413dfbd-3c8b-42ac-8536-c0331fb7a397-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-rmtvx\" (UID: \"5413dfbd-3c8b-42ac-8536-c0331fb7a397\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rmtvx" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.362819 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.363818 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5413dfbd-3c8b-42ac-8536-c0331fb7a397-serving-cert\") pod \"controller-manager-879f6c89f-rmtvx\" (UID: \"5413dfbd-3c8b-42ac-8536-c0331fb7a397\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rmtvx" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.364183 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/3ce01e2e-eb84-47e0-a3c9-d4b0044f79a5-machine-approver-tls\") pod \"machine-approver-56656f9798-sxxzq\" (UID: \"3ce01e2e-eb84-47e0-a3c9-d4b0044f79a5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sxxzq" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.365384 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b619039f-82c2-49ca-9319-75d0883a5a7d-serving-cert\") pod \"apiserver-7bbb656c7d-xsm2s\" (UID: \"b619039f-82c2-49ca-9319-75d0883a5a7d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.367001 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8b821731-b65f-45df-b510-de915e573be1-node-pullsecrets\") pod \"apiserver-76f77b778f-pvqd2\" (UID: \"8b821731-b65f-45df-b510-de915e573be1\") " pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.367224 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8b821731-b65f-45df-b510-de915e573be1-audit-dir\") pod \"apiserver-76f77b778f-pvqd2\" (UID: \"8b821731-b65f-45df-b510-de915e573be1\") " pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.369968 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/26e29bc4-82c6-47c1-8ffa-5e35e4b1c595-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-9bq59\" (UID: \"26e29bc4-82c6-47c1-8ffa-5e35e4b1c595\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9bq59" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.370411 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dd97957f-e7df-48c2-b23f-b1101b786766-trusted-ca-bundle\") pod \"console-f9d7485db-mbmx2\" (UID: \"dd97957f-e7df-48c2-b23f-b1101b786766\") " pod="openshift-console/console-f9d7485db-mbmx2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.370475 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/358a66f5-3192-4b20-ba92-37c8a8544209-available-featuregates\") pod \"openshift-config-operator-7777fb866f-c2xs7\" (UID: \"358a66f5-3192-4b20-ba92-37c8a8544209\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-c2xs7" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.370597 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ce01e2e-eb84-47e0-a3c9-d4b0044f79a5-config\") pod \"machine-approver-56656f9798-sxxzq\" (UID: \"3ce01e2e-eb84-47e0-a3c9-d4b0044f79a5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sxxzq" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.370649 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lkx85"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.370743 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb22f225-d744-4357-a091-d21d76a31e42-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-wdn7d\" (UID: \"bb22f225-d744-4357-a091-d21d76a31e42\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wdn7d" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.371423 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7cfd6b45-6486-4b59-9a79-21773426f02b-serving-cert\") pod \"authentication-operator-69f744f599-dnwh7\" (UID: \"7cfd6b45-6486-4b59-9a79-21773426f02b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dnwh7" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.370351 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3ce01e2e-eb84-47e0-a3c9-d4b0044f79a5-auth-proxy-config\") pod \"machine-approver-56656f9798-sxxzq\" (UID: \"3ce01e2e-eb84-47e0-a3c9-d4b0044f79a5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sxxzq" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.371546 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a0fafb6-8df4-457c-9651-1d40cc9742b3-config\") pod \"service-ca-operator-777779d784-5tvj9\" (UID: \"7a0fafb6-8df4-457c-9651-1d40cc9742b3\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5tvj9" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.372305 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/dd97957f-e7df-48c2-b23f-b1101b786766-console-oauth-config\") pod \"console-f9d7485db-mbmx2\" (UID: \"dd97957f-e7df-48c2-b23f-b1101b786766\") " pod="openshift-console/console-f9d7485db-mbmx2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.372584 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b2jt8"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.372714 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5413dfbd-3c8b-42ac-8536-c0331fb7a397-config\") pod \"controller-manager-879f6c89f-rmtvx\" (UID: \"5413dfbd-3c8b-42ac-8536-c0331fb7a397\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rmtvx" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.373952 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b619039f-82c2-49ca-9319-75d0883a5a7d-audit-policies\") pod \"apiserver-7bbb656c7d-xsm2s\" (UID: \"b619039f-82c2-49ca-9319-75d0883a5a7d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.374176 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.375109 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-8vr2l"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.377028 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/358a66f5-3192-4b20-ba92-37c8a8544209-serving-cert\") pod \"openshift-config-operator-7777fb866f-c2xs7\" (UID: \"358a66f5-3192-4b20-ba92-37c8a8544209\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-c2xs7" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.376310 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b619039f-82c2-49ca-9319-75d0883a5a7d-encryption-config\") pod \"apiserver-7bbb656c7d-xsm2s\" (UID: \"b619039f-82c2-49ca-9319-75d0883a5a7d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.376547 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/017380dd-ff61-49ce-8603-00daaeef7d5f-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-4zttp\" (UID: \"017380dd-ff61-49ce-8603-00daaeef7d5f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4zttp" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.376576 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d9232dc2-d71f-41b1-864a-82603860af69-proxy-tls\") pod \"machine-config-controller-84d6567774-bjl94\" (UID: \"d9232dc2-d71f-41b1-864a-82603860af69\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bjl94" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.377073 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-8vt69"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.375846 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7a0fafb6-8df4-457c-9651-1d40cc9742b3-serving-cert\") pod \"service-ca-operator-777779d784-5tvj9\" (UID: \"7a0fafb6-8df4-457c-9651-1d40cc9742b3\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5tvj9" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.377908 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vlz75"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.378021 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-8vt69" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.379084 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/25f66de5-fdbd-44dc-911d-976d5c2bbfd3-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-b2jt8\" (UID: \"25f66de5-fdbd-44dc-911d-976d5c2bbfd3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b2jt8" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.379566 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b821731-b65f-45df-b510-de915e573be1-serving-cert\") pod \"apiserver-76f77b778f-pvqd2\" (UID: \"8b821731-b65f-45df-b510-de915e573be1\") " pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.379947 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.380440 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/25f66de5-fdbd-44dc-911d-976d5c2bbfd3-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-b2jt8\" (UID: \"25f66de5-fdbd-44dc-911d-976d5c2bbfd3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b2jt8" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.380798 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.381014 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vjbjr"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.380732 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7cfd6b45-6486-4b59-9a79-21773426f02b-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-dnwh7\" (UID: \"7cfd6b45-6486-4b59-9a79-21773426f02b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dnwh7" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.381271 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/dd97957f-e7df-48c2-b23f-b1101b786766-console-serving-cert\") pod \"console-f9d7485db-mbmx2\" (UID: \"dd97957f-e7df-48c2-b23f-b1101b786766\") " pod="openshift-console/console-f9d7485db-mbmx2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.381951 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gj88f"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.382700 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.382804 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8b821731-b65f-45df-b510-de915e573be1-encryption-config\") pod \"apiserver-76f77b778f-pvqd2\" (UID: \"8b821731-b65f-45df-b510-de915e573be1\") " pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.382976 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.383207 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4zsfb"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.383540 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5add632c-b0b2-490d-a0f6-d0193c176134-metrics-tls\") pod \"dns-operator-744455d44c-q2bqm\" (UID: \"5add632c-b0b2-490d-a0f6-d0193c176134\") " pod="openshift-dns-operator/dns-operator-744455d44c-q2bqm" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.384062 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.386701 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.387465 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.388080 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4c1d0e3-c3c8-47f3-ba79-d8aa252b3572-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9xwkx\" (UID: \"e4c1d0e3-c3c8-47f3-ba79-d8aa252b3572\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9xwkx" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.388456 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-zqrgn"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.389067 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/8b821731-b65f-45df-b510-de915e573be1-audit\") pod \"apiserver-76f77b778f-pvqd2\" (UID: \"8b821731-b65f-45df-b510-de915e573be1\") " pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.389113 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b619039f-82c2-49ca-9319-75d0883a5a7d-etcd-client\") pod \"apiserver-7bbb656c7d-xsm2s\" (UID: \"b619039f-82c2-49ca-9319-75d0883a5a7d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.389138 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8b821731-b65f-45df-b510-de915e573be1-etcd-client\") pod \"apiserver-76f77b778f-pvqd2\" (UID: \"8b821731-b65f-45df-b510-de915e573be1\") " pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.389289 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8b821731-b65f-45df-b510-de915e573be1-etcd-serving-ca\") pod \"apiserver-76f77b778f-pvqd2\" (UID: \"8b821731-b65f-45df-b510-de915e573be1\") " pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.389673 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b821731-b65f-45df-b510-de915e573be1-config\") pod \"apiserver-76f77b778f-pvqd2\" (UID: \"8b821731-b65f-45df-b510-de915e573be1\") " pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.391544 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-dfr9q"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.393126 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.393279 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-dkk9l"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.393805 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.395261 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-bmb9v"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.395349 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-dkk9l" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.395817 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401305-j5rqf"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.397307 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73-serving-cert\") pod \"route-controller-manager-6576b87f9c-wv6k6\" (UID: \"ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wv6k6" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.397705 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-twgvd"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.399032 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5vx95"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.400502 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-j6d5l"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.401764 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9xwkx"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.402907 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mb6kd"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.406369 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-5tvj9"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.406427 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-dkk9l"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.407220 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-mlxwh"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.408301 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-jqmtw"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.409415 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-5cmt4"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.411715 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dk89m"] Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.418190 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.434378 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.454569 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.481503 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.494900 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.534146 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.553934 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.576041 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.594351 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.614125 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.634640 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.655102 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.674144 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.694753 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.714551 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.734465 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.755956 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.774007 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.794095 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.814541 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.837065 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.854252 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.874154 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.894070 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.913835 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.933675 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.954301 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.974791 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 25 13:48:03 crc kubenswrapper[4702]: I1125 13:48:03.994520 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.013325 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.034647 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.054179 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.073758 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.094146 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.115874 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.134229 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.153541 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.174399 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.194018 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.213987 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.233555 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.254262 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.273671 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.293653 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.312005 4702 request.go:700] Waited for 1.007408486s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-etcd-operator/secrets?fieldSelector=metadata.name%3Detcd-client&limit=500&resourceVersion=0 Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.313566 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.334174 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.356221 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.373976 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.393847 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.414707 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.434762 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.454227 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.474628 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.493804 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.514820 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.540869 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.553006 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.574073 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.594666 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.622759 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.633587 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.655335 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.675672 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.694560 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.714599 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.735286 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.755649 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.775337 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.794579 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.800260 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.800289 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.800439 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.800508 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.815060 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.834392 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.856379 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.875222 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.894473 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.915134 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.934916 4702 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.955197 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 25 13:48:04 crc kubenswrapper[4702]: I1125 13:48:04.994570 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.013733 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.034516 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.070134 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsd8v\" (UniqueName: \"kubernetes.io/projected/26e29bc4-82c6-47c1-8ffa-5e35e4b1c595-kube-api-access-nsd8v\") pod \"machine-api-operator-5694c8668f-9bq59\" (UID: \"26e29bc4-82c6-47c1-8ffa-5e35e4b1c595\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9bq59" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.103365 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gv77\" (UniqueName: \"kubernetes.io/projected/25f66de5-fdbd-44dc-911d-976d5c2bbfd3-kube-api-access-9gv77\") pod \"cluster-image-registry-operator-dc59b4c8b-b2jt8\" (UID: \"25f66de5-fdbd-44dc-911d-976d5c2bbfd3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b2jt8" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.113329 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8cqf\" (UniqueName: \"kubernetes.io/projected/ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73-kube-api-access-q8cqf\") pod \"route-controller-manager-6576b87f9c-wv6k6\" (UID: \"ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wv6k6" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.138533 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cd67b\" (UniqueName: \"kubernetes.io/projected/ab12e899-20e3-4755-a537-3e54c5761af6-kube-api-access-cd67b\") pod \"oauth-openshift-558db77b4-dfr9q\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.159649 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/25f66de5-fdbd-44dc-911d-976d5c2bbfd3-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-b2jt8\" (UID: \"25f66de5-fdbd-44dc-911d-976d5c2bbfd3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b2jt8" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.173434 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rjll\" (UniqueName: \"kubernetes.io/projected/017380dd-ff61-49ce-8603-00daaeef7d5f-kube-api-access-8rjll\") pod \"cluster-samples-operator-665b6dd947-4zttp\" (UID: \"017380dd-ff61-49ce-8603-00daaeef7d5f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4zttp" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.197510 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hn5lj\" (UniqueName: \"kubernetes.io/projected/7cfd6b45-6486-4b59-9a79-21773426f02b-kube-api-access-hn5lj\") pod \"authentication-operator-69f744f599-dnwh7\" (UID: \"7cfd6b45-6486-4b59-9a79-21773426f02b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dnwh7" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.215844 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkcwj\" (UniqueName: \"kubernetes.io/projected/5add632c-b0b2-490d-a0f6-d0193c176134-kube-api-access-qkcwj\") pod \"dns-operator-744455d44c-q2bqm\" (UID: \"5add632c-b0b2-490d-a0f6-d0193c176134\") " pod="openshift-dns-operator/dns-operator-744455d44c-q2bqm" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.228772 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e4c1d0e3-c3c8-47f3-ba79-d8aa252b3572-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9xwkx\" (UID: \"e4c1d0e3-c3c8-47f3-ba79-d8aa252b3572\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9xwkx" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.229789 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.238992 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-q2bqm" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.245788 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9xwkx" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.256269 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtjbt\" (UniqueName: \"kubernetes.io/projected/bb22f225-d744-4357-a091-d21d76a31e42-kube-api-access-qtjbt\") pod \"openshift-apiserver-operator-796bbdcf4f-wdn7d\" (UID: \"bb22f225-d744-4357-a091-d21d76a31e42\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wdn7d" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.268417 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wv6k6" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.275109 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxpsk\" (UniqueName: \"kubernetes.io/projected/8b821731-b65f-45df-b510-de915e573be1-kube-api-access-fxpsk\") pod \"apiserver-76f77b778f-pvqd2\" (UID: \"8b821731-b65f-45df-b510-de915e573be1\") " pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.288328 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4zttp" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.295651 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbs8d\" (UniqueName: \"kubernetes.io/projected/3ce01e2e-eb84-47e0-a3c9-d4b0044f79a5-kube-api-access-nbs8d\") pod \"machine-approver-56656f9798-sxxzq\" (UID: \"3ce01e2e-eb84-47e0-a3c9-d4b0044f79a5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sxxzq" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.312048 4702 request.go:700] Waited for 1.94607783s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-console/serviceaccounts/console/token Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.315050 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g96tq\" (UniqueName: \"kubernetes.io/projected/7a0fafb6-8df4-457c-9651-1d40cc9742b3-kube-api-access-g96tq\") pod \"service-ca-operator-777779d784-5tvj9\" (UID: \"7a0fafb6-8df4-457c-9651-1d40cc9742b3\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5tvj9" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.322895 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wdn7d" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.334811 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsgkj\" (UniqueName: \"kubernetes.io/projected/dd97957f-e7df-48c2-b23f-b1101b786766-kube-api-access-bsgkj\") pod \"console-f9d7485db-mbmx2\" (UID: \"dd97957f-e7df-48c2-b23f-b1101b786766\") " pod="openshift-console/console-f9d7485db-mbmx2" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.346917 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-9bq59" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.356315 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9l8v\" (UniqueName: \"kubernetes.io/projected/358a66f5-3192-4b20-ba92-37c8a8544209-kube-api-access-r9l8v\") pod \"openshift-config-operator-7777fb866f-c2xs7\" (UID: \"358a66f5-3192-4b20-ba92-37c8a8544209\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-c2xs7" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.367465 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.375129 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ccsm\" (UniqueName: \"kubernetes.io/projected/5413dfbd-3c8b-42ac-8536-c0331fb7a397-kube-api-access-7ccsm\") pod \"controller-manager-879f6c89f-rmtvx\" (UID: \"5413dfbd-3c8b-42ac-8536-c0331fb7a397\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rmtvx" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.401507 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rmtvx" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.402342 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jp8wv\" (UniqueName: \"kubernetes.io/projected/b619039f-82c2-49ca-9319-75d0883a5a7d-kube-api-access-jp8wv\") pod \"apiserver-7bbb656c7d-xsm2s\" (UID: \"b619039f-82c2-49ca-9319-75d0883a5a7d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.409071 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mbmx2" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.423801 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b2jt8" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.430407 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rpfs\" (UniqueName: \"kubernetes.io/projected/d9232dc2-d71f-41b1-864a-82603860af69-kube-api-access-7rpfs\") pod \"machine-config-controller-84d6567774-bjl94\" (UID: \"d9232dc2-d71f-41b1-864a-82603860af69\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bjl94" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.432555 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqh98\" (UniqueName: \"kubernetes.io/projected/ce097f6c-2894-4417-b66b-0f30d91f3889-kube-api-access-zqh98\") pod \"downloads-7954f5f757-8j9xk\" (UID: \"ce097f6c-2894-4417-b66b-0f30d91f3889\") " pod="openshift-console/downloads-7954f5f757-8j9xk" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.434076 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.440670 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-dnwh7" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.454443 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.456049 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.462189 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sxxzq" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.474148 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.493561 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.519979 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.534665 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.554683 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-5tvj9" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.556720 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.561476 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bjl94" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.597977 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.614362 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.628627 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-c2xs7" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.637073 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.637656 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-8j9xk" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.654499 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.674563 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.694203 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.788991 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a70d6c59-6b85-4950-8e30-1938da7a01e8-trusted-ca\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789026 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7de9c8c4-69ff-478f-9a13-ea39ca9a0db5-signing-key\") pod \"service-ca-9c57cc56f-j6d5l\" (UID: \"7de9c8c4-69ff-478f-9a13-ea39ca9a0db5\") " pod="openshift-service-ca/service-ca-9c57cc56f-j6d5l" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789045 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f74394e4-d2c6-4211-be82-2f1c1de9c536-etcd-client\") pod \"etcd-operator-b45778765-j7l2t\" (UID: \"f74394e4-d2c6-4211-be82-2f1c1de9c536\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j7l2t" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789061 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8cb1a228-426f-41cc-a15f-01dfb84a35cd-config-volume\") pod \"collect-profiles-29401305-j5rqf\" (UID: \"8cb1a228-426f-41cc-a15f-01dfb84a35cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401305-j5rqf" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789090 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a70d6c59-6b85-4950-8e30-1938da7a01e8-ca-trust-extracted\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789107 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a70d6c59-6b85-4950-8e30-1938da7a01e8-installation-pull-secrets\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789142 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6d7fc28a-71f2-415c-a99e-76e4ab7fbb70-metrics-certs\") pod \"router-default-5444994796-lmq49\" (UID: \"6d7fc28a-71f2-415c-a99e-76e4ab7fbb70\") " pod="openshift-ingress/router-default-5444994796-lmq49" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789158 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5829q\" (UniqueName: \"kubernetes.io/projected/a70d6c59-6b85-4950-8e30-1938da7a01e8-kube-api-access-5829q\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789185 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ec7a64ba-34c9-4e31-8e2b-1991cba108d0-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lkx85\" (UID: \"ec7a64ba-34c9-4e31-8e2b-1991cba108d0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lkx85" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789202 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/8f55435c-3542-4faf-997d-8a7dad39ce50-csi-data-dir\") pod \"csi-hostpathplugin-5cmt4\" (UID: \"8f55435c-3542-4faf-997d-8a7dad39ce50\") " pod="hostpath-provisioner/csi-hostpathplugin-5cmt4" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789265 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/24dafe4a-2dce-4e4a-8a2c-d5653b326861-trusted-ca\") pod \"ingress-operator-5b745b69d9-mlxwh\" (UID: \"24dafe4a-2dce-4e4a-8a2c-d5653b326861\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mlxwh" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789314 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/8f55435c-3542-4faf-997d-8a7dad39ce50-registration-dir\") pod \"csi-hostpathplugin-5cmt4\" (UID: \"8f55435c-3542-4faf-997d-8a7dad39ce50\") " pod="hostpath-provisioner/csi-hostpathplugin-5cmt4" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789328 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/17279013-fed3-475f-b960-3b9e41116da8-webhook-cert\") pod \"packageserver-d55dfcdfc-t5vtr\" (UID: \"17279013-fed3-475f-b960-3b9e41116da8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t5vtr" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789355 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0c290048-0c68-42fe-bd32-985795e837eb-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-xszrd\" (UID: \"0c290048-0c68-42fe-bd32-985795e837eb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xszrd" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789370 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvcqp\" (UniqueName: \"kubernetes.io/projected/17279013-fed3-475f-b960-3b9e41116da8-kube-api-access-dvcqp\") pod \"packageserver-d55dfcdfc-t5vtr\" (UID: \"17279013-fed3-475f-b960-3b9e41116da8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t5vtr" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789385 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6d7fc28a-71f2-415c-a99e-76e4ab7fbb70-service-ca-bundle\") pod \"router-default-5444994796-lmq49\" (UID: \"6d7fc28a-71f2-415c-a99e-76e4ab7fbb70\") " pod="openshift-ingress/router-default-5444994796-lmq49" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789411 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c290048-0c68-42fe-bd32-985795e837eb-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-xszrd\" (UID: \"0c290048-0c68-42fe-bd32-985795e837eb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xszrd" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789453 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/edba4243-475a-49b8-8d1d-183164f35a54-proxy-tls\") pod \"machine-config-operator-74547568cd-8vr2l\" (UID: \"edba4243-475a-49b8-8d1d-183164f35a54\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8vr2l" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789499 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbf919e2-9e7d-4e22-a04d-e838be932151-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-mb6kd\" (UID: \"cbf919e2-9e7d-4e22-a04d-e838be932151\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mb6kd" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789533 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a70d6c59-6b85-4950-8e30-1938da7a01e8-registry-tls\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789547 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/8f55435c-3542-4faf-997d-8a7dad39ce50-mountpoint-dir\") pod \"csi-hostpathplugin-5cmt4\" (UID: \"8f55435c-3542-4faf-997d-8a7dad39ce50\") " pod="hostpath-provisioner/csi-hostpathplugin-5cmt4" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789609 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2190c8e2-2c7d-47fd-bb13-32ff0020c655-serving-cert\") pod \"console-operator-58897d9998-7twx2\" (UID: \"2190c8e2-2c7d-47fd-bb13-32ff0020c655\") " pod="openshift-console-operator/console-operator-58897d9998-7twx2" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789637 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpnkc\" (UniqueName: \"kubernetes.io/projected/138d7b9f-9cae-436f-a06e-e3ccd0b9168d-kube-api-access-lpnkc\") pod \"control-plane-machine-set-operator-78cbb6b69f-vjbjr\" (UID: \"138d7b9f-9cae-436f-a06e-e3ccd0b9168d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vjbjr" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789652 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/17279013-fed3-475f-b960-3b9e41116da8-tmpfs\") pod \"packageserver-d55dfcdfc-t5vtr\" (UID: \"17279013-fed3-475f-b960-3b9e41116da8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t5vtr" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789667 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a70d6c59-6b85-4950-8e30-1938da7a01e8-bound-sa-token\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789695 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/6d7fc28a-71f2-415c-a99e-76e4ab7fbb70-default-certificate\") pod \"router-default-5444994796-lmq49\" (UID: \"6d7fc28a-71f2-415c-a99e-76e4ab7fbb70\") " pod="openshift-ingress/router-default-5444994796-lmq49" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789710 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/edba4243-475a-49b8-8d1d-183164f35a54-auth-proxy-config\") pod \"machine-config-operator-74547568cd-8vr2l\" (UID: \"edba4243-475a-49b8-8d1d-183164f35a54\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8vr2l" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789727 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f8caa35c-4401-482b-8c5e-2927fbb50c4d-profile-collector-cert\") pod \"olm-operator-6b444d44fb-dk89m\" (UID: \"f8caa35c-4401-482b-8c5e-2927fbb50c4d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dk89m" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789755 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppvrd\" (UniqueName: \"kubernetes.io/projected/8f55435c-3542-4faf-997d-8a7dad39ce50-kube-api-access-ppvrd\") pod \"csi-hostpathplugin-5cmt4\" (UID: \"8f55435c-3542-4faf-997d-8a7dad39ce50\") " pod="hostpath-provisioner/csi-hostpathplugin-5cmt4" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789771 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmvwl\" (UniqueName: \"kubernetes.io/projected/0c290048-0c68-42fe-bd32-985795e837eb-kube-api-access-gmvwl\") pod \"openshift-controller-manager-operator-756b6f6bc6-xszrd\" (UID: \"0c290048-0c68-42fe-bd32-985795e837eb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xszrd" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789787 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgbkn\" (UniqueName: \"kubernetes.io/projected/8cb1a228-426f-41cc-a15f-01dfb84a35cd-kube-api-access-jgbkn\") pod \"collect-profiles-29401305-j5rqf\" (UID: \"8cb1a228-426f-41cc-a15f-01dfb84a35cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401305-j5rqf" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789804 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9b430b6f-d294-4e5a-9410-fcdc8fca501d-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-bmb9v\" (UID: \"9b430b6f-d294-4e5a-9410-fcdc8fca501d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-bmb9v" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789821 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a70d6c59-6b85-4950-8e30-1938da7a01e8-registry-certificates\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789849 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgr9k\" (UniqueName: \"kubernetes.io/projected/f8caa35c-4401-482b-8c5e-2927fbb50c4d-kube-api-access-vgr9k\") pod \"olm-operator-6b444d44fb-dk89m\" (UID: \"f8caa35c-4401-482b-8c5e-2927fbb50c4d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dk89m" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789876 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/17279013-fed3-475f-b960-3b9e41116da8-apiservice-cert\") pod \"packageserver-d55dfcdfc-t5vtr\" (UID: \"17279013-fed3-475f-b960-3b9e41116da8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t5vtr" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789901 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qk8s8\" (UniqueName: \"kubernetes.io/projected/9b430b6f-d294-4e5a-9410-fcdc8fca501d-kube-api-access-qk8s8\") pod \"multus-admission-controller-857f4d67dd-bmb9v\" (UID: \"9b430b6f-d294-4e5a-9410-fcdc8fca501d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-bmb9v" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789931 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwrfb\" (UniqueName: \"kubernetes.io/projected/92fb9b54-8df1-40ac-8c8c-66d12416bc74-kube-api-access-hwrfb\") pod \"package-server-manager-789f6589d5-4zsfb\" (UID: \"92fb9b54-8df1-40ac-8c8c-66d12416bc74\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4zsfb" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789947 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f74394e4-d2c6-4211-be82-2f1c1de9c536-etcd-service-ca\") pod \"etcd-operator-b45778765-j7l2t\" (UID: \"f74394e4-d2c6-4211-be82-2f1c1de9c536\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j7l2t" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.789963 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwn47\" (UniqueName: \"kubernetes.io/projected/f74394e4-d2c6-4211-be82-2f1c1de9c536-kube-api-access-jwn47\") pod \"etcd-operator-b45778765-j7l2t\" (UID: \"f74394e4-d2c6-4211-be82-2f1c1de9c536\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j7l2t" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790007 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/92fb9b54-8df1-40ac-8c8c-66d12416bc74-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-4zsfb\" (UID: \"92fb9b54-8df1-40ac-8c8c-66d12416bc74\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4zsfb" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790044 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/6d7fc28a-71f2-415c-a99e-76e4ab7fbb70-stats-auth\") pod \"router-default-5444994796-lmq49\" (UID: \"6d7fc28a-71f2-415c-a99e-76e4ab7fbb70\") " pod="openshift-ingress/router-default-5444994796-lmq49" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790073 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d41be55e-0563-404d-b2b0-d642b675910e-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-5vx95\" (UID: \"d41be55e-0563-404d-b2b0-d642b675910e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5vx95" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790101 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d41be55e-0563-404d-b2b0-d642b675910e-config\") pod \"kube-apiserver-operator-766d6c64bb-5vx95\" (UID: \"d41be55e-0563-404d-b2b0-d642b675910e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5vx95" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790117 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/24dafe4a-2dce-4e4a-8a2c-d5653b326861-metrics-tls\") pod \"ingress-operator-5b745b69d9-mlxwh\" (UID: \"24dafe4a-2dce-4e4a-8a2c-d5653b326861\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mlxwh" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790131 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/8f55435c-3542-4faf-997d-8a7dad39ce50-plugins-dir\") pod \"csi-hostpathplugin-5cmt4\" (UID: \"8f55435c-3542-4faf-997d-8a7dad39ce50\") " pod="hostpath-provisioner/csi-hostpathplugin-5cmt4" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790158 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/138d7b9f-9cae-436f-a06e-e3ccd0b9168d-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-vjbjr\" (UID: \"138d7b9f-9cae-436f-a06e-e3ccd0b9168d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vjbjr" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790185 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxhtc\" (UniqueName: \"kubernetes.io/projected/30dd0e87-b5d2-4ef3-896f-aaaaf64300ad-kube-api-access-jxhtc\") pod \"dns-default-jqmtw\" (UID: \"30dd0e87-b5d2-4ef3-896f-aaaaf64300ad\") " pod="openshift-dns/dns-default-jqmtw" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790210 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mt6dk\" (UniqueName: \"kubernetes.io/projected/7de9c8c4-69ff-478f-9a13-ea39ca9a0db5-kube-api-access-mt6dk\") pod \"service-ca-9c57cc56f-j6d5l\" (UID: \"7de9c8c4-69ff-478f-9a13-ea39ca9a0db5\") " pod="openshift-service-ca/service-ca-9c57cc56f-j6d5l" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790269 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/24dafe4a-2dce-4e4a-8a2c-d5653b326861-bound-sa-token\") pod \"ingress-operator-5b745b69d9-mlxwh\" (UID: \"24dafe4a-2dce-4e4a-8a2c-d5653b326861\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mlxwh" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790289 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8cb1a228-426f-41cc-a15f-01dfb84a35cd-secret-volume\") pod \"collect-profiles-29401305-j5rqf\" (UID: \"8cb1a228-426f-41cc-a15f-01dfb84a35cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401305-j5rqf" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790308 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/89437c7c-7034-44f8-bb77-000a12163744-srv-cert\") pod \"catalog-operator-68c6474976-gj88f\" (UID: \"89437c7c-7034-44f8-bb77-000a12163744\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gj88f" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790325 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/89437c7c-7034-44f8-bb77-000a12163744-profile-collector-cert\") pod \"catalog-operator-68c6474976-gj88f\" (UID: \"89437c7c-7034-44f8-bb77-000a12163744\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gj88f" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790351 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbf919e2-9e7d-4e22-a04d-e838be932151-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-mb6kd\" (UID: \"cbf919e2-9e7d-4e22-a04d-e838be932151\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mb6kd" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790366 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f74394e4-d2c6-4211-be82-2f1c1de9c536-config\") pod \"etcd-operator-b45778765-j7l2t\" (UID: \"f74394e4-d2c6-4211-be82-2f1c1de9c536\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j7l2t" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790381 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vz97\" (UniqueName: \"kubernetes.io/projected/f51cb34e-c816-48dd-98d9-63ccc57222d7-kube-api-access-7vz97\") pod \"migrator-59844c95c7-twgvd\" (UID: \"f51cb34e-c816-48dd-98d9-63ccc57222d7\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-twgvd" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790403 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0389edfa-9a86-48df-a9a6-26e2b4d092c4-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-zqrgn\" (UID: \"0389edfa-9a86-48df-a9a6-26e2b4d092c4\") " pod="openshift-marketplace/marketplace-operator-79b997595-zqrgn" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790529 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4d6w\" (UniqueName: \"kubernetes.io/projected/edba4243-475a-49b8-8d1d-183164f35a54-kube-api-access-t4d6w\") pod \"machine-config-operator-74547568cd-8vr2l\" (UID: \"edba4243-475a-49b8-8d1d-183164f35a54\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8vr2l" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790574 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ec7a64ba-34c9-4e31-8e2b-1991cba108d0-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lkx85\" (UID: \"ec7a64ba-34c9-4e31-8e2b-1991cba108d0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lkx85" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790591 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/edba4243-475a-49b8-8d1d-183164f35a54-images\") pod \"machine-config-operator-74547568cd-8vr2l\" (UID: \"edba4243-475a-49b8-8d1d-183164f35a54\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8vr2l" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790607 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hntpv\" (UniqueName: \"kubernetes.io/projected/89437c7c-7034-44f8-bb77-000a12163744-kube-api-access-hntpv\") pod \"catalog-operator-68c6474976-gj88f\" (UID: \"89437c7c-7034-44f8-bb77-000a12163744\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gj88f" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790634 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2190c8e2-2c7d-47fd-bb13-32ff0020c655-config\") pod \"console-operator-58897d9998-7twx2\" (UID: \"2190c8e2-2c7d-47fd-bb13-32ff0020c655\") " pod="openshift-console-operator/console-operator-58897d9998-7twx2" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790658 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7de9c8c4-69ff-478f-9a13-ea39ca9a0db5-signing-cabundle\") pod \"service-ca-9c57cc56f-j6d5l\" (UID: \"7de9c8c4-69ff-478f-9a13-ea39ca9a0db5\") " pod="openshift-service-ca/service-ca-9c57cc56f-j6d5l" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790691 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2190c8e2-2c7d-47fd-bb13-32ff0020c655-trusted-ca\") pod \"console-operator-58897d9998-7twx2\" (UID: \"2190c8e2-2c7d-47fd-bb13-32ff0020c655\") " pod="openshift-console-operator/console-operator-58897d9998-7twx2" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790707 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnhzh\" (UniqueName: \"kubernetes.io/projected/6d7fc28a-71f2-415c-a99e-76e4ab7fbb70-kube-api-access-nnhzh\") pod \"router-default-5444994796-lmq49\" (UID: \"6d7fc28a-71f2-415c-a99e-76e4ab7fbb70\") " pod="openshift-ingress/router-default-5444994796-lmq49" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790771 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/30dd0e87-b5d2-4ef3-896f-aaaaf64300ad-config-volume\") pod \"dns-default-jqmtw\" (UID: \"30dd0e87-b5d2-4ef3-896f-aaaaf64300ad\") " pod="openshift-dns/dns-default-jqmtw" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790787 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/30dd0e87-b5d2-4ef3-896f-aaaaf64300ad-metrics-tls\") pod \"dns-default-jqmtw\" (UID: \"30dd0e87-b5d2-4ef3-896f-aaaaf64300ad\") " pod="openshift-dns/dns-default-jqmtw" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790840 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f74394e4-d2c6-4211-be82-2f1c1de9c536-etcd-ca\") pod \"etcd-operator-b45778765-j7l2t\" (UID: \"f74394e4-d2c6-4211-be82-2f1c1de9c536\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j7l2t" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790855 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d41be55e-0563-404d-b2b0-d642b675910e-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-5vx95\" (UID: \"d41be55e-0563-404d-b2b0-d642b675910e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5vx95" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790910 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8flvp\" (UniqueName: \"kubernetes.io/projected/cbf919e2-9e7d-4e22-a04d-e838be932151-kube-api-access-8flvp\") pod \"kube-storage-version-migrator-operator-b67b599dd-mb6kd\" (UID: \"cbf919e2-9e7d-4e22-a04d-e838be932151\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mb6kd" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790928 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwcsz\" (UniqueName: \"kubernetes.io/projected/24dafe4a-2dce-4e4a-8a2c-d5653b326861-kube-api-access-dwcsz\") pod \"ingress-operator-5b745b69d9-mlxwh\" (UID: \"24dafe4a-2dce-4e4a-8a2c-d5653b326861\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mlxwh" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790944 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0389edfa-9a86-48df-a9a6-26e2b4d092c4-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-zqrgn\" (UID: \"0389edfa-9a86-48df-a9a6-26e2b4d092c4\") " pod="openshift-marketplace/marketplace-operator-79b997595-zqrgn" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790960 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vgk7\" (UniqueName: \"kubernetes.io/projected/0389edfa-9a86-48df-a9a6-26e2b4d092c4-kube-api-access-2vgk7\") pod \"marketplace-operator-79b997595-zqrgn\" (UID: \"0389edfa-9a86-48df-a9a6-26e2b4d092c4\") " pod="openshift-marketplace/marketplace-operator-79b997595-zqrgn" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.790978 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f74394e4-d2c6-4211-be82-2f1c1de9c536-serving-cert\") pod \"etcd-operator-b45778765-j7l2t\" (UID: \"f74394e4-d2c6-4211-be82-2f1c1de9c536\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j7l2t" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.791004 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/8f55435c-3542-4faf-997d-8a7dad39ce50-socket-dir\") pod \"csi-hostpathplugin-5cmt4\" (UID: \"8f55435c-3542-4faf-997d-8a7dad39ce50\") " pod="hostpath-provisioner/csi-hostpathplugin-5cmt4" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.791020 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtwqv\" (UniqueName: \"kubernetes.io/projected/2190c8e2-2c7d-47fd-bb13-32ff0020c655-kube-api-access-jtwqv\") pod \"console-operator-58897d9998-7twx2\" (UID: \"2190c8e2-2c7d-47fd-bb13-32ff0020c655\") " pod="openshift-console-operator/console-operator-58897d9998-7twx2" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.791036 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f8caa35c-4401-482b-8c5e-2927fbb50c4d-srv-cert\") pod \"olm-operator-6b444d44fb-dk89m\" (UID: \"f8caa35c-4401-482b-8c5e-2927fbb50c4d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dk89m" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.791085 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec7a64ba-34c9-4e31-8e2b-1991cba108d0-config\") pod \"kube-controller-manager-operator-78b949d7b-lkx85\" (UID: \"ec7a64ba-34c9-4e31-8e2b-1991cba108d0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lkx85" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.791127 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:05 crc kubenswrapper[4702]: E1125 13:48:05.791674 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:06.291659773 +0000 UTC m=+95.454310223 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.882726 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-q2bqm"] Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.883635 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-dfr9q"] Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.883726 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9xwkx"] Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.893912 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:05 crc kubenswrapper[4702]: E1125 13:48:05.895731 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:06.394544035 +0000 UTC m=+95.557194485 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.898988 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4d6w\" (UniqueName: \"kubernetes.io/projected/edba4243-475a-49b8-8d1d-183164f35a54-kube-api-access-t4d6w\") pod \"machine-config-operator-74547568cd-8vr2l\" (UID: \"edba4243-475a-49b8-8d1d-183164f35a54\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8vr2l" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.899059 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ec7a64ba-34c9-4e31-8e2b-1991cba108d0-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lkx85\" (UID: \"ec7a64ba-34c9-4e31-8e2b-1991cba108d0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lkx85" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.899100 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/edba4243-475a-49b8-8d1d-183164f35a54-images\") pod \"machine-config-operator-74547568cd-8vr2l\" (UID: \"edba4243-475a-49b8-8d1d-183164f35a54\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8vr2l" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.899127 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hntpv\" (UniqueName: \"kubernetes.io/projected/89437c7c-7034-44f8-bb77-000a12163744-kube-api-access-hntpv\") pod \"catalog-operator-68c6474976-gj88f\" (UID: \"89437c7c-7034-44f8-bb77-000a12163744\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gj88f" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.899154 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2190c8e2-2c7d-47fd-bb13-32ff0020c655-config\") pod \"console-operator-58897d9998-7twx2\" (UID: \"2190c8e2-2c7d-47fd-bb13-32ff0020c655\") " pod="openshift-console-operator/console-operator-58897d9998-7twx2" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.899179 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7de9c8c4-69ff-478f-9a13-ea39ca9a0db5-signing-cabundle\") pod \"service-ca-9c57cc56f-j6d5l\" (UID: \"7de9c8c4-69ff-478f-9a13-ea39ca9a0db5\") " pod="openshift-service-ca/service-ca-9c57cc56f-j6d5l" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.899210 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2190c8e2-2c7d-47fd-bb13-32ff0020c655-trusted-ca\") pod \"console-operator-58897d9998-7twx2\" (UID: \"2190c8e2-2c7d-47fd-bb13-32ff0020c655\") " pod="openshift-console-operator/console-operator-58897d9998-7twx2" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.899249 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnhzh\" (UniqueName: \"kubernetes.io/projected/6d7fc28a-71f2-415c-a99e-76e4ab7fbb70-kube-api-access-nnhzh\") pod \"router-default-5444994796-lmq49\" (UID: \"6d7fc28a-71f2-415c-a99e-76e4ab7fbb70\") " pod="openshift-ingress/router-default-5444994796-lmq49" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.899291 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/30dd0e87-b5d2-4ef3-896f-aaaaf64300ad-config-volume\") pod \"dns-default-jqmtw\" (UID: \"30dd0e87-b5d2-4ef3-896f-aaaaf64300ad\") " pod="openshift-dns/dns-default-jqmtw" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.899324 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/30dd0e87-b5d2-4ef3-896f-aaaaf64300ad-metrics-tls\") pod \"dns-default-jqmtw\" (UID: \"30dd0e87-b5d2-4ef3-896f-aaaaf64300ad\") " pod="openshift-dns/dns-default-jqmtw" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.899370 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f74394e4-d2c6-4211-be82-2f1c1de9c536-etcd-ca\") pod \"etcd-operator-b45778765-j7l2t\" (UID: \"f74394e4-d2c6-4211-be82-2f1c1de9c536\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j7l2t" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.899407 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d41be55e-0563-404d-b2b0-d642b675910e-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-5vx95\" (UID: \"d41be55e-0563-404d-b2b0-d642b675910e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5vx95" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.899441 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8flvp\" (UniqueName: \"kubernetes.io/projected/cbf919e2-9e7d-4e22-a04d-e838be932151-kube-api-access-8flvp\") pod \"kube-storage-version-migrator-operator-b67b599dd-mb6kd\" (UID: \"cbf919e2-9e7d-4e22-a04d-e838be932151\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mb6kd" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.899476 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwcsz\" (UniqueName: \"kubernetes.io/projected/24dafe4a-2dce-4e4a-8a2c-d5653b326861-kube-api-access-dwcsz\") pod \"ingress-operator-5b745b69d9-mlxwh\" (UID: \"24dafe4a-2dce-4e4a-8a2c-d5653b326861\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mlxwh" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.899499 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0389edfa-9a86-48df-a9a6-26e2b4d092c4-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-zqrgn\" (UID: \"0389edfa-9a86-48df-a9a6-26e2b4d092c4\") " pod="openshift-marketplace/marketplace-operator-79b997595-zqrgn" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.899519 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vgk7\" (UniqueName: \"kubernetes.io/projected/0389edfa-9a86-48df-a9a6-26e2b4d092c4-kube-api-access-2vgk7\") pod \"marketplace-operator-79b997595-zqrgn\" (UID: \"0389edfa-9a86-48df-a9a6-26e2b4d092c4\") " pod="openshift-marketplace/marketplace-operator-79b997595-zqrgn" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.899539 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f74394e4-d2c6-4211-be82-2f1c1de9c536-serving-cert\") pod \"etcd-operator-b45778765-j7l2t\" (UID: \"f74394e4-d2c6-4211-be82-2f1c1de9c536\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j7l2t" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.899565 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/8f55435c-3542-4faf-997d-8a7dad39ce50-socket-dir\") pod \"csi-hostpathplugin-5cmt4\" (UID: \"8f55435c-3542-4faf-997d-8a7dad39ce50\") " pod="hostpath-provisioner/csi-hostpathplugin-5cmt4" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.899586 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtwqv\" (UniqueName: \"kubernetes.io/projected/2190c8e2-2c7d-47fd-bb13-32ff0020c655-kube-api-access-jtwqv\") pod \"console-operator-58897d9998-7twx2\" (UID: \"2190c8e2-2c7d-47fd-bb13-32ff0020c655\") " pod="openshift-console-operator/console-operator-58897d9998-7twx2" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.899609 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f8caa35c-4401-482b-8c5e-2927fbb50c4d-srv-cert\") pod \"olm-operator-6b444d44fb-dk89m\" (UID: \"f8caa35c-4401-482b-8c5e-2927fbb50c4d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dk89m" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.899646 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec7a64ba-34c9-4e31-8e2b-1991cba108d0-config\") pod \"kube-controller-manager-operator-78b949d7b-lkx85\" (UID: \"ec7a64ba-34c9-4e31-8e2b-1991cba108d0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lkx85" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.899689 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.899716 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a70d6c59-6b85-4950-8e30-1938da7a01e8-trusted-ca\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.899737 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7de9c8c4-69ff-478f-9a13-ea39ca9a0db5-signing-key\") pod \"service-ca-9c57cc56f-j6d5l\" (UID: \"7de9c8c4-69ff-478f-9a13-ea39ca9a0db5\") " pod="openshift-service-ca/service-ca-9c57cc56f-j6d5l" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.899757 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f74394e4-d2c6-4211-be82-2f1c1de9c536-etcd-client\") pod \"etcd-operator-b45778765-j7l2t\" (UID: \"f74394e4-d2c6-4211-be82-2f1c1de9c536\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j7l2t" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.899777 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8cb1a228-426f-41cc-a15f-01dfb84a35cd-config-volume\") pod \"collect-profiles-29401305-j5rqf\" (UID: \"8cb1a228-426f-41cc-a15f-01dfb84a35cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401305-j5rqf" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.899843 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/f961361a-3a5a-49be-b34a-06857d960b39-node-bootstrap-token\") pod \"machine-config-server-8vt69\" (UID: \"f961361a-3a5a-49be-b34a-06857d960b39\") " pod="openshift-machine-config-operator/machine-config-server-8vt69" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.899869 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a70d6c59-6b85-4950-8e30-1938da7a01e8-ca-trust-extracted\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.899888 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a70d6c59-6b85-4950-8e30-1938da7a01e8-installation-pull-secrets\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.899929 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6d7fc28a-71f2-415c-a99e-76e4ab7fbb70-metrics-certs\") pod \"router-default-5444994796-lmq49\" (UID: \"6d7fc28a-71f2-415c-a99e-76e4ab7fbb70\") " pod="openshift-ingress/router-default-5444994796-lmq49" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.899952 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5829q\" (UniqueName: \"kubernetes.io/projected/a70d6c59-6b85-4950-8e30-1938da7a01e8-kube-api-access-5829q\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.899978 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ec7a64ba-34c9-4e31-8e2b-1991cba108d0-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lkx85\" (UID: \"ec7a64ba-34c9-4e31-8e2b-1991cba108d0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lkx85" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900004 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/90b61112-2c79-4b92-930f-a182a569c0ae-cert\") pod \"ingress-canary-dkk9l\" (UID: \"90b61112-2c79-4b92-930f-a182a569c0ae\") " pod="openshift-ingress-canary/ingress-canary-dkk9l" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900027 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/8f55435c-3542-4faf-997d-8a7dad39ce50-csi-data-dir\") pod \"csi-hostpathplugin-5cmt4\" (UID: \"8f55435c-3542-4faf-997d-8a7dad39ce50\") " pod="hostpath-provisioner/csi-hostpathplugin-5cmt4" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900060 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/24dafe4a-2dce-4e4a-8a2c-d5653b326861-trusted-ca\") pod \"ingress-operator-5b745b69d9-mlxwh\" (UID: \"24dafe4a-2dce-4e4a-8a2c-d5653b326861\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mlxwh" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900092 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/8f55435c-3542-4faf-997d-8a7dad39ce50-registration-dir\") pod \"csi-hostpathplugin-5cmt4\" (UID: \"8f55435c-3542-4faf-997d-8a7dad39ce50\") " pod="hostpath-provisioner/csi-hostpathplugin-5cmt4" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900115 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/17279013-fed3-475f-b960-3b9e41116da8-webhook-cert\") pod \"packageserver-d55dfcdfc-t5vtr\" (UID: \"17279013-fed3-475f-b960-3b9e41116da8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t5vtr" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900137 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0c290048-0c68-42fe-bd32-985795e837eb-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-xszrd\" (UID: \"0c290048-0c68-42fe-bd32-985795e837eb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xszrd" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900161 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvcqp\" (UniqueName: \"kubernetes.io/projected/17279013-fed3-475f-b960-3b9e41116da8-kube-api-access-dvcqp\") pod \"packageserver-d55dfcdfc-t5vtr\" (UID: \"17279013-fed3-475f-b960-3b9e41116da8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t5vtr" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900185 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6d7fc28a-71f2-415c-a99e-76e4ab7fbb70-service-ca-bundle\") pod \"router-default-5444994796-lmq49\" (UID: \"6d7fc28a-71f2-415c-a99e-76e4ab7fbb70\") " pod="openshift-ingress/router-default-5444994796-lmq49" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900207 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c290048-0c68-42fe-bd32-985795e837eb-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-xszrd\" (UID: \"0c290048-0c68-42fe-bd32-985795e837eb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xszrd" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900253 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/edba4243-475a-49b8-8d1d-183164f35a54-proxy-tls\") pod \"machine-config-operator-74547568cd-8vr2l\" (UID: \"edba4243-475a-49b8-8d1d-183164f35a54\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8vr2l" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900277 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bz2p7\" (UniqueName: \"kubernetes.io/projected/90b61112-2c79-4b92-930f-a182a569c0ae-kube-api-access-bz2p7\") pod \"ingress-canary-dkk9l\" (UID: \"90b61112-2c79-4b92-930f-a182a569c0ae\") " pod="openshift-ingress-canary/ingress-canary-dkk9l" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900308 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbf919e2-9e7d-4e22-a04d-e838be932151-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-mb6kd\" (UID: \"cbf919e2-9e7d-4e22-a04d-e838be932151\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mb6kd" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900333 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a70d6c59-6b85-4950-8e30-1938da7a01e8-registry-tls\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900357 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/8f55435c-3542-4faf-997d-8a7dad39ce50-mountpoint-dir\") pod \"csi-hostpathplugin-5cmt4\" (UID: \"8f55435c-3542-4faf-997d-8a7dad39ce50\") " pod="hostpath-provisioner/csi-hostpathplugin-5cmt4" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900391 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2190c8e2-2c7d-47fd-bb13-32ff0020c655-serving-cert\") pod \"console-operator-58897d9998-7twx2\" (UID: \"2190c8e2-2c7d-47fd-bb13-32ff0020c655\") " pod="openshift-console-operator/console-operator-58897d9998-7twx2" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900424 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpnkc\" (UniqueName: \"kubernetes.io/projected/138d7b9f-9cae-436f-a06e-e3ccd0b9168d-kube-api-access-lpnkc\") pod \"control-plane-machine-set-operator-78cbb6b69f-vjbjr\" (UID: \"138d7b9f-9cae-436f-a06e-e3ccd0b9168d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vjbjr" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900448 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/17279013-fed3-475f-b960-3b9e41116da8-tmpfs\") pod \"packageserver-d55dfcdfc-t5vtr\" (UID: \"17279013-fed3-475f-b960-3b9e41116da8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t5vtr" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900470 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a70d6c59-6b85-4950-8e30-1938da7a01e8-bound-sa-token\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900502 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/6d7fc28a-71f2-415c-a99e-76e4ab7fbb70-default-certificate\") pod \"router-default-5444994796-lmq49\" (UID: \"6d7fc28a-71f2-415c-a99e-76e4ab7fbb70\") " pod="openshift-ingress/router-default-5444994796-lmq49" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900522 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/edba4243-475a-49b8-8d1d-183164f35a54-auth-proxy-config\") pod \"machine-config-operator-74547568cd-8vr2l\" (UID: \"edba4243-475a-49b8-8d1d-183164f35a54\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8vr2l" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900546 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f8caa35c-4401-482b-8c5e-2927fbb50c4d-profile-collector-cert\") pod \"olm-operator-6b444d44fb-dk89m\" (UID: \"f8caa35c-4401-482b-8c5e-2927fbb50c4d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dk89m" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900573 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppvrd\" (UniqueName: \"kubernetes.io/projected/8f55435c-3542-4faf-997d-8a7dad39ce50-kube-api-access-ppvrd\") pod \"csi-hostpathplugin-5cmt4\" (UID: \"8f55435c-3542-4faf-997d-8a7dad39ce50\") " pod="hostpath-provisioner/csi-hostpathplugin-5cmt4" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900599 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmvwl\" (UniqueName: \"kubernetes.io/projected/0c290048-0c68-42fe-bd32-985795e837eb-kube-api-access-gmvwl\") pod \"openshift-controller-manager-operator-756b6f6bc6-xszrd\" (UID: \"0c290048-0c68-42fe-bd32-985795e837eb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xszrd" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900618 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgbkn\" (UniqueName: \"kubernetes.io/projected/8cb1a228-426f-41cc-a15f-01dfb84a35cd-kube-api-access-jgbkn\") pod \"collect-profiles-29401305-j5rqf\" (UID: \"8cb1a228-426f-41cc-a15f-01dfb84a35cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401305-j5rqf" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900641 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9b430b6f-d294-4e5a-9410-fcdc8fca501d-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-bmb9v\" (UID: \"9b430b6f-d294-4e5a-9410-fcdc8fca501d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-bmb9v" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900664 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a70d6c59-6b85-4950-8e30-1938da7a01e8-registry-certificates\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900689 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgr9k\" (UniqueName: \"kubernetes.io/projected/f8caa35c-4401-482b-8c5e-2927fbb50c4d-kube-api-access-vgr9k\") pod \"olm-operator-6b444d44fb-dk89m\" (UID: \"f8caa35c-4401-482b-8c5e-2927fbb50c4d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dk89m" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900714 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/17279013-fed3-475f-b960-3b9e41116da8-apiservice-cert\") pod \"packageserver-d55dfcdfc-t5vtr\" (UID: \"17279013-fed3-475f-b960-3b9e41116da8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t5vtr" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900738 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qk8s8\" (UniqueName: \"kubernetes.io/projected/9b430b6f-d294-4e5a-9410-fcdc8fca501d-kube-api-access-qk8s8\") pod \"multus-admission-controller-857f4d67dd-bmb9v\" (UID: \"9b430b6f-d294-4e5a-9410-fcdc8fca501d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-bmb9v" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900766 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwrfb\" (UniqueName: \"kubernetes.io/projected/92fb9b54-8df1-40ac-8c8c-66d12416bc74-kube-api-access-hwrfb\") pod \"package-server-manager-789f6589d5-4zsfb\" (UID: \"92fb9b54-8df1-40ac-8c8c-66d12416bc74\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4zsfb" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900791 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f74394e4-d2c6-4211-be82-2f1c1de9c536-etcd-service-ca\") pod \"etcd-operator-b45778765-j7l2t\" (UID: \"f74394e4-d2c6-4211-be82-2f1c1de9c536\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j7l2t" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900821 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwn47\" (UniqueName: \"kubernetes.io/projected/f74394e4-d2c6-4211-be82-2f1c1de9c536-kube-api-access-jwn47\") pod \"etcd-operator-b45778765-j7l2t\" (UID: \"f74394e4-d2c6-4211-be82-2f1c1de9c536\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j7l2t" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900873 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/92fb9b54-8df1-40ac-8c8c-66d12416bc74-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-4zsfb\" (UID: \"92fb9b54-8df1-40ac-8c8c-66d12416bc74\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4zsfb" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900896 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/f961361a-3a5a-49be-b34a-06857d960b39-certs\") pod \"machine-config-server-8vt69\" (UID: \"f961361a-3a5a-49be-b34a-06857d960b39\") " pod="openshift-machine-config-operator/machine-config-server-8vt69" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900923 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/6d7fc28a-71f2-415c-a99e-76e4ab7fbb70-stats-auth\") pod \"router-default-5444994796-lmq49\" (UID: \"6d7fc28a-71f2-415c-a99e-76e4ab7fbb70\") " pod="openshift-ingress/router-default-5444994796-lmq49" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900950 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d41be55e-0563-404d-b2b0-d642b675910e-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-5vx95\" (UID: \"d41be55e-0563-404d-b2b0-d642b675910e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5vx95" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900976 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d41be55e-0563-404d-b2b0-d642b675910e-config\") pod \"kube-apiserver-operator-766d6c64bb-5vx95\" (UID: \"d41be55e-0563-404d-b2b0-d642b675910e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5vx95" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.900995 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/24dafe4a-2dce-4e4a-8a2c-d5653b326861-metrics-tls\") pod \"ingress-operator-5b745b69d9-mlxwh\" (UID: \"24dafe4a-2dce-4e4a-8a2c-d5653b326861\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mlxwh" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.901019 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/8f55435c-3542-4faf-997d-8a7dad39ce50-plugins-dir\") pod \"csi-hostpathplugin-5cmt4\" (UID: \"8f55435c-3542-4faf-997d-8a7dad39ce50\") " pod="hostpath-provisioner/csi-hostpathplugin-5cmt4" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.901047 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/138d7b9f-9cae-436f-a06e-e3ccd0b9168d-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-vjbjr\" (UID: \"138d7b9f-9cae-436f-a06e-e3ccd0b9168d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vjbjr" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.901074 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxhtc\" (UniqueName: \"kubernetes.io/projected/30dd0e87-b5d2-4ef3-896f-aaaaf64300ad-kube-api-access-jxhtc\") pod \"dns-default-jqmtw\" (UID: \"30dd0e87-b5d2-4ef3-896f-aaaaf64300ad\") " pod="openshift-dns/dns-default-jqmtw" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.901100 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hn9hb\" (UniqueName: \"kubernetes.io/projected/f961361a-3a5a-49be-b34a-06857d960b39-kube-api-access-hn9hb\") pod \"machine-config-server-8vt69\" (UID: \"f961361a-3a5a-49be-b34a-06857d960b39\") " pod="openshift-machine-config-operator/machine-config-server-8vt69" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.901131 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mt6dk\" (UniqueName: \"kubernetes.io/projected/7de9c8c4-69ff-478f-9a13-ea39ca9a0db5-kube-api-access-mt6dk\") pod \"service-ca-9c57cc56f-j6d5l\" (UID: \"7de9c8c4-69ff-478f-9a13-ea39ca9a0db5\") " pod="openshift-service-ca/service-ca-9c57cc56f-j6d5l" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.901158 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/24dafe4a-2dce-4e4a-8a2c-d5653b326861-bound-sa-token\") pod \"ingress-operator-5b745b69d9-mlxwh\" (UID: \"24dafe4a-2dce-4e4a-8a2c-d5653b326861\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mlxwh" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.901181 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8cb1a228-426f-41cc-a15f-01dfb84a35cd-secret-volume\") pod \"collect-profiles-29401305-j5rqf\" (UID: \"8cb1a228-426f-41cc-a15f-01dfb84a35cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401305-j5rqf" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.901205 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/89437c7c-7034-44f8-bb77-000a12163744-srv-cert\") pod \"catalog-operator-68c6474976-gj88f\" (UID: \"89437c7c-7034-44f8-bb77-000a12163744\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gj88f" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.901226 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/89437c7c-7034-44f8-bb77-000a12163744-profile-collector-cert\") pod \"catalog-operator-68c6474976-gj88f\" (UID: \"89437c7c-7034-44f8-bb77-000a12163744\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gj88f" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.901346 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbf919e2-9e7d-4e22-a04d-e838be932151-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-mb6kd\" (UID: \"cbf919e2-9e7d-4e22-a04d-e838be932151\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mb6kd" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.901370 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f74394e4-d2c6-4211-be82-2f1c1de9c536-config\") pod \"etcd-operator-b45778765-j7l2t\" (UID: \"f74394e4-d2c6-4211-be82-2f1c1de9c536\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j7l2t" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.901400 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vz97\" (UniqueName: \"kubernetes.io/projected/f51cb34e-c816-48dd-98d9-63ccc57222d7-kube-api-access-7vz97\") pod \"migrator-59844c95c7-twgvd\" (UID: \"f51cb34e-c816-48dd-98d9-63ccc57222d7\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-twgvd" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.901424 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0389edfa-9a86-48df-a9a6-26e2b4d092c4-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-zqrgn\" (UID: \"0389edfa-9a86-48df-a9a6-26e2b4d092c4\") " pod="openshift-marketplace/marketplace-operator-79b997595-zqrgn" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.903208 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbf919e2-9e7d-4e22-a04d-e838be932151-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-mb6kd\" (UID: \"cbf919e2-9e7d-4e22-a04d-e838be932151\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mb6kd" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.905186 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec7a64ba-34c9-4e31-8e2b-1991cba108d0-config\") pod \"kube-controller-manager-operator-78b949d7b-lkx85\" (UID: \"ec7a64ba-34c9-4e31-8e2b-1991cba108d0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lkx85" Nov 25 13:48:05 crc kubenswrapper[4702]: E1125 13:48:05.905895 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:06.405871604 +0000 UTC m=+95.568522054 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.906004 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/edba4243-475a-49b8-8d1d-183164f35a54-images\") pod \"machine-config-operator-74547568cd-8vr2l\" (UID: \"edba4243-475a-49b8-8d1d-183164f35a54\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8vr2l" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.906123 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f74394e4-d2c6-4211-be82-2f1c1de9c536-etcd-service-ca\") pod \"etcd-operator-b45778765-j7l2t\" (UID: \"f74394e4-d2c6-4211-be82-2f1c1de9c536\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j7l2t" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.906738 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2190c8e2-2c7d-47fd-bb13-32ff0020c655-config\") pod \"console-operator-58897d9998-7twx2\" (UID: \"2190c8e2-2c7d-47fd-bb13-32ff0020c655\") " pod="openshift-console-operator/console-operator-58897d9998-7twx2" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.906766 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/edba4243-475a-49b8-8d1d-183164f35a54-proxy-tls\") pod \"machine-config-operator-74547568cd-8vr2l\" (UID: \"edba4243-475a-49b8-8d1d-183164f35a54\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8vr2l" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.907493 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/92fb9b54-8df1-40ac-8c8c-66d12416bc74-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-4zsfb\" (UID: \"92fb9b54-8df1-40ac-8c8c-66d12416bc74\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4zsfb" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.907588 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7de9c8c4-69ff-478f-9a13-ea39ca9a0db5-signing-cabundle\") pod \"service-ca-9c57cc56f-j6d5l\" (UID: \"7de9c8c4-69ff-478f-9a13-ea39ca9a0db5\") " pod="openshift-service-ca/service-ca-9c57cc56f-j6d5l" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.912666 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a70d6c59-6b85-4950-8e30-1938da7a01e8-registry-certificates\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.908962 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0389edfa-9a86-48df-a9a6-26e2b4d092c4-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-zqrgn\" (UID: \"0389edfa-9a86-48df-a9a6-26e2b4d092c4\") " pod="openshift-marketplace/marketplace-operator-79b997595-zqrgn" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.909089 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2190c8e2-2c7d-47fd-bb13-32ff0020c655-trusted-ca\") pod \"console-operator-58897d9998-7twx2\" (UID: \"2190c8e2-2c7d-47fd-bb13-32ff0020c655\") " pod="openshift-console-operator/console-operator-58897d9998-7twx2" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.910074 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/8f55435c-3542-4faf-997d-8a7dad39ce50-mountpoint-dir\") pod \"csi-hostpathplugin-5cmt4\" (UID: \"8f55435c-3542-4faf-997d-8a7dad39ce50\") " pod="hostpath-provisioner/csi-hostpathplugin-5cmt4" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.911471 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/8f55435c-3542-4faf-997d-8a7dad39ce50-plugins-dir\") pod \"csi-hostpathplugin-5cmt4\" (UID: \"8f55435c-3542-4faf-997d-8a7dad39ce50\") " pod="hostpath-provisioner/csi-hostpathplugin-5cmt4" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.912485 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/8f55435c-3542-4faf-997d-8a7dad39ce50-socket-dir\") pod \"csi-hostpathplugin-5cmt4\" (UID: \"8f55435c-3542-4faf-997d-8a7dad39ce50\") " pod="hostpath-provisioner/csi-hostpathplugin-5cmt4" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.908140 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a70d6c59-6b85-4950-8e30-1938da7a01e8-trusted-ca\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.913326 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/17279013-fed3-475f-b960-3b9e41116da8-tmpfs\") pod \"packageserver-d55dfcdfc-t5vtr\" (UID: \"17279013-fed3-475f-b960-3b9e41116da8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t5vtr" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.913785 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f8caa35c-4401-482b-8c5e-2927fbb50c4d-profile-collector-cert\") pod \"olm-operator-6b444d44fb-dk89m\" (UID: \"f8caa35c-4401-482b-8c5e-2927fbb50c4d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dk89m" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.913851 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/30dd0e87-b5d2-4ef3-896f-aaaaf64300ad-config-volume\") pod \"dns-default-jqmtw\" (UID: \"30dd0e87-b5d2-4ef3-896f-aaaaf64300ad\") " pod="openshift-dns/dns-default-jqmtw" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.913934 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d41be55e-0563-404d-b2b0-d642b675910e-config\") pod \"kube-apiserver-operator-766d6c64bb-5vx95\" (UID: \"d41be55e-0563-404d-b2b0-d642b675910e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5vx95" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.914281 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8cb1a228-426f-41cc-a15f-01dfb84a35cd-config-volume\") pod \"collect-profiles-29401305-j5rqf\" (UID: \"8cb1a228-426f-41cc-a15f-01dfb84a35cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401305-j5rqf" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.914387 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d41be55e-0563-404d-b2b0-d642b675910e-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-5vx95\" (UID: \"d41be55e-0563-404d-b2b0-d642b675910e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5vx95" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.914610 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0389edfa-9a86-48df-a9a6-26e2b4d092c4-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-zqrgn\" (UID: \"0389edfa-9a86-48df-a9a6-26e2b4d092c4\") " pod="openshift-marketplace/marketplace-operator-79b997595-zqrgn" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.914615 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/6d7fc28a-71f2-415c-a99e-76e4ab7fbb70-stats-auth\") pod \"router-default-5444994796-lmq49\" (UID: \"6d7fc28a-71f2-415c-a99e-76e4ab7fbb70\") " pod="openshift-ingress/router-default-5444994796-lmq49" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.914770 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a70d6c59-6b85-4950-8e30-1938da7a01e8-ca-trust-extracted\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.914799 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/8f55435c-3542-4faf-997d-8a7dad39ce50-registration-dir\") pod \"csi-hostpathplugin-5cmt4\" (UID: \"8f55435c-3542-4faf-997d-8a7dad39ce50\") " pod="hostpath-provisioner/csi-hostpathplugin-5cmt4" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.915701 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a70d6c59-6b85-4950-8e30-1938da7a01e8-registry-tls\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.915877 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/8f55435c-3542-4faf-997d-8a7dad39ce50-csi-data-dir\") pod \"csi-hostpathplugin-5cmt4\" (UID: \"8f55435c-3542-4faf-997d-8a7dad39ce50\") " pod="hostpath-provisioner/csi-hostpathplugin-5cmt4" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.916369 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f74394e4-d2c6-4211-be82-2f1c1de9c536-config\") pod \"etcd-operator-b45778765-j7l2t\" (UID: \"f74394e4-d2c6-4211-be82-2f1c1de9c536\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j7l2t" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.916892 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c290048-0c68-42fe-bd32-985795e837eb-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-xszrd\" (UID: \"0c290048-0c68-42fe-bd32-985795e837eb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xszrd" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.917524 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6d7fc28a-71f2-415c-a99e-76e4ab7fbb70-service-ca-bundle\") pod \"router-default-5444994796-lmq49\" (UID: \"6d7fc28a-71f2-415c-a99e-76e4ab7fbb70\") " pod="openshift-ingress/router-default-5444994796-lmq49" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.917564 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/edba4243-475a-49b8-8d1d-183164f35a54-auth-proxy-config\") pod \"machine-config-operator-74547568cd-8vr2l\" (UID: \"edba4243-475a-49b8-8d1d-183164f35a54\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8vr2l" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.917582 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/24dafe4a-2dce-4e4a-8a2c-d5653b326861-trusted-ca\") pod \"ingress-operator-5b745b69d9-mlxwh\" (UID: \"24dafe4a-2dce-4e4a-8a2c-d5653b326861\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mlxwh" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.918276 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wv6k6"] Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.918658 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f74394e4-d2c6-4211-be82-2f1c1de9c536-etcd-ca\") pod \"etcd-operator-b45778765-j7l2t\" (UID: \"f74394e4-d2c6-4211-be82-2f1c1de9c536\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j7l2t" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.918812 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/30dd0e87-b5d2-4ef3-896f-aaaaf64300ad-metrics-tls\") pod \"dns-default-jqmtw\" (UID: \"30dd0e87-b5d2-4ef3-896f-aaaaf64300ad\") " pod="openshift-dns/dns-default-jqmtw" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.919464 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0c290048-0c68-42fe-bd32-985795e837eb-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-xszrd\" (UID: \"0c290048-0c68-42fe-bd32-985795e837eb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xszrd" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.919882 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9b430b6f-d294-4e5a-9410-fcdc8fca501d-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-bmb9v\" (UID: \"9b430b6f-d294-4e5a-9410-fcdc8fca501d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-bmb9v" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.921718 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6d7fc28a-71f2-415c-a99e-76e4ab7fbb70-metrics-certs\") pod \"router-default-5444994796-lmq49\" (UID: \"6d7fc28a-71f2-415c-a99e-76e4ab7fbb70\") " pod="openshift-ingress/router-default-5444994796-lmq49" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.922055 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/24dafe4a-2dce-4e4a-8a2c-d5653b326861-metrics-tls\") pod \"ingress-operator-5b745b69d9-mlxwh\" (UID: \"24dafe4a-2dce-4e4a-8a2c-d5653b326861\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mlxwh" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.926788 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f74394e4-d2c6-4211-be82-2f1c1de9c536-serving-cert\") pod \"etcd-operator-b45778765-j7l2t\" (UID: \"f74394e4-d2c6-4211-be82-2f1c1de9c536\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j7l2t" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.927534 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/138d7b9f-9cae-436f-a06e-e3ccd0b9168d-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-vjbjr\" (UID: \"138d7b9f-9cae-436f-a06e-e3ccd0b9168d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vjbjr" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.929312 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ec7a64ba-34c9-4e31-8e2b-1991cba108d0-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lkx85\" (UID: \"ec7a64ba-34c9-4e31-8e2b-1991cba108d0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lkx85" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.929839 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f74394e4-d2c6-4211-be82-2f1c1de9c536-etcd-client\") pod \"etcd-operator-b45778765-j7l2t\" (UID: \"f74394e4-d2c6-4211-be82-2f1c1de9c536\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j7l2t" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.930211 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8cb1a228-426f-41cc-a15f-01dfb84a35cd-secret-volume\") pod \"collect-profiles-29401305-j5rqf\" (UID: \"8cb1a228-426f-41cc-a15f-01dfb84a35cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401305-j5rqf" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.930414 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/89437c7c-7034-44f8-bb77-000a12163744-srv-cert\") pod \"catalog-operator-68c6474976-gj88f\" (UID: \"89437c7c-7034-44f8-bb77-000a12163744\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gj88f" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.930604 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/6d7fc28a-71f2-415c-a99e-76e4ab7fbb70-default-certificate\") pod \"router-default-5444994796-lmq49\" (UID: \"6d7fc28a-71f2-415c-a99e-76e4ab7fbb70\") " pod="openshift-ingress/router-default-5444994796-lmq49" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.930958 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/89437c7c-7034-44f8-bb77-000a12163744-profile-collector-cert\") pod \"catalog-operator-68c6474976-gj88f\" (UID: \"89437c7c-7034-44f8-bb77-000a12163744\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gj88f" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.932571 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a70d6c59-6b85-4950-8e30-1938da7a01e8-installation-pull-secrets\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.935854 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbf919e2-9e7d-4e22-a04d-e838be932151-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-mb6kd\" (UID: \"cbf919e2-9e7d-4e22-a04d-e838be932151\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mb6kd" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.935968 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f8caa35c-4401-482b-8c5e-2927fbb50c4d-srv-cert\") pod \"olm-operator-6b444d44fb-dk89m\" (UID: \"f8caa35c-4401-482b-8c5e-2927fbb50c4d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dk89m" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.936079 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/17279013-fed3-475f-b960-3b9e41116da8-apiservice-cert\") pod \"packageserver-d55dfcdfc-t5vtr\" (UID: \"17279013-fed3-475f-b960-3b9e41116da8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t5vtr" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.936267 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2190c8e2-2c7d-47fd-bb13-32ff0020c655-serving-cert\") pod \"console-operator-58897d9998-7twx2\" (UID: \"2190c8e2-2c7d-47fd-bb13-32ff0020c655\") " pod="openshift-console-operator/console-operator-58897d9998-7twx2" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.943393 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7de9c8c4-69ff-478f-9a13-ea39ca9a0db5-signing-key\") pod \"service-ca-9c57cc56f-j6d5l\" (UID: \"7de9c8c4-69ff-478f-9a13-ea39ca9a0db5\") " pod="openshift-service-ca/service-ca-9c57cc56f-j6d5l" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.954776 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4d6w\" (UniqueName: \"kubernetes.io/projected/edba4243-475a-49b8-8d1d-183164f35a54-kube-api-access-t4d6w\") pod \"machine-config-operator-74547568cd-8vr2l\" (UID: \"edba4243-475a-49b8-8d1d-183164f35a54\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8vr2l" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.955269 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/17279013-fed3-475f-b960-3b9e41116da8-webhook-cert\") pod \"packageserver-d55dfcdfc-t5vtr\" (UID: \"17279013-fed3-475f-b960-3b9e41116da8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t5vtr" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.983210 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwn47\" (UniqueName: \"kubernetes.io/projected/f74394e4-d2c6-4211-be82-2f1c1de9c536-kube-api-access-jwn47\") pod \"etcd-operator-b45778765-j7l2t\" (UID: \"f74394e4-d2c6-4211-be82-2f1c1de9c536\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j7l2t" Nov 25 13:48:05 crc kubenswrapper[4702]: I1125 13:48:05.994091 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppvrd\" (UniqueName: \"kubernetes.io/projected/8f55435c-3542-4faf-997d-8a7dad39ce50-kube-api-access-ppvrd\") pod \"csi-hostpathplugin-5cmt4\" (UID: \"8f55435c-3542-4faf-997d-8a7dad39ce50\") " pod="hostpath-provisioner/csi-hostpathplugin-5cmt4" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.001778 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-j7l2t" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.002920 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.003188 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/f961361a-3a5a-49be-b34a-06857d960b39-certs\") pod \"machine-config-server-8vt69\" (UID: \"f961361a-3a5a-49be-b34a-06857d960b39\") " pod="openshift-machine-config-operator/machine-config-server-8vt69" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.003223 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hn9hb\" (UniqueName: \"kubernetes.io/projected/f961361a-3a5a-49be-b34a-06857d960b39-kube-api-access-hn9hb\") pod \"machine-config-server-8vt69\" (UID: \"f961361a-3a5a-49be-b34a-06857d960b39\") " pod="openshift-machine-config-operator/machine-config-server-8vt69" Nov 25 13:48:06 crc kubenswrapper[4702]: E1125 13:48:06.003384 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:06.503359042 +0000 UTC m=+95.666009492 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.003640 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/f961361a-3a5a-49be-b34a-06857d960b39-node-bootstrap-token\") pod \"machine-config-server-8vt69\" (UID: \"f961361a-3a5a-49be-b34a-06857d960b39\") " pod="openshift-machine-config-operator/machine-config-server-8vt69" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.003720 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/90b61112-2c79-4b92-930f-a182a569c0ae-cert\") pod \"ingress-canary-dkk9l\" (UID: \"90b61112-2c79-4b92-930f-a182a569c0ae\") " pod="openshift-ingress-canary/ingress-canary-dkk9l" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.003788 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bz2p7\" (UniqueName: \"kubernetes.io/projected/90b61112-2c79-4b92-930f-a182a569c0ae-kube-api-access-bz2p7\") pod \"ingress-canary-dkk9l\" (UID: \"90b61112-2c79-4b92-930f-a182a569c0ae\") " pod="openshift-ingress-canary/ingress-canary-dkk9l" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.009466 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/f961361a-3a5a-49be-b34a-06857d960b39-node-bootstrap-token\") pod \"machine-config-server-8vt69\" (UID: \"f961361a-3a5a-49be-b34a-06857d960b39\") " pod="openshift-machine-config-operator/machine-config-server-8vt69" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.009649 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/90b61112-2c79-4b92-930f-a182a569c0ae-cert\") pod \"ingress-canary-dkk9l\" (UID: \"90b61112-2c79-4b92-930f-a182a569c0ae\") " pod="openshift-ingress-canary/ingress-canary-dkk9l" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.009788 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/f961361a-3a5a-49be-b34a-06857d960b39-certs\") pod \"machine-config-server-8vt69\" (UID: \"f961361a-3a5a-49be-b34a-06857d960b39\") " pod="openshift-machine-config-operator/machine-config-server-8vt69" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.026665 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmvwl\" (UniqueName: \"kubernetes.io/projected/0c290048-0c68-42fe-bd32-985795e837eb-kube-api-access-gmvwl\") pod \"openshift-controller-manager-operator-756b6f6bc6-xszrd\" (UID: \"0c290048-0c68-42fe-bd32-985795e837eb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xszrd" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.029253 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgbkn\" (UniqueName: \"kubernetes.io/projected/8cb1a228-426f-41cc-a15f-01dfb84a35cd-kube-api-access-jgbkn\") pod \"collect-profiles-29401305-j5rqf\" (UID: \"8cb1a228-426f-41cc-a15f-01dfb84a35cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401305-j5rqf" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.054101 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ec7a64ba-34c9-4e31-8e2b-1991cba108d0-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lkx85\" (UID: \"ec7a64ba-34c9-4e31-8e2b-1991cba108d0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lkx85" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.081213 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hntpv\" (UniqueName: \"kubernetes.io/projected/89437c7c-7034-44f8-bb77-000a12163744-kube-api-access-hntpv\") pod \"catalog-operator-68c6474976-gj88f\" (UID: \"89437c7c-7034-44f8-bb77-000a12163744\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gj88f" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.100414 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8flvp\" (UniqueName: \"kubernetes.io/projected/cbf919e2-9e7d-4e22-a04d-e838be932151-kube-api-access-8flvp\") pod \"kube-storage-version-migrator-operator-b67b599dd-mb6kd\" (UID: \"cbf919e2-9e7d-4e22-a04d-e838be932151\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mb6kd" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.100485 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-5cmt4" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.107186 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:06 crc kubenswrapper[4702]: E1125 13:48:06.107644 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:06.607625988 +0000 UTC m=+95.770276438 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.110419 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwcsz\" (UniqueName: \"kubernetes.io/projected/24dafe4a-2dce-4e4a-8a2c-d5653b326861-kube-api-access-dwcsz\") pod \"ingress-operator-5b745b69d9-mlxwh\" (UID: \"24dafe4a-2dce-4e4a-8a2c-d5653b326861\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mlxwh" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.133584 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vgk7\" (UniqueName: \"kubernetes.io/projected/0389edfa-9a86-48df-a9a6-26e2b4d092c4-kube-api-access-2vgk7\") pod \"marketplace-operator-79b997595-zqrgn\" (UID: \"0389edfa-9a86-48df-a9a6-26e2b4d092c4\") " pod="openshift-marketplace/marketplace-operator-79b997595-zqrgn" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.160059 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnhzh\" (UniqueName: \"kubernetes.io/projected/6d7fc28a-71f2-415c-a99e-76e4ab7fbb70-kube-api-access-nnhzh\") pod \"router-default-5444994796-lmq49\" (UID: \"6d7fc28a-71f2-415c-a99e-76e4ab7fbb70\") " pod="openshift-ingress/router-default-5444994796-lmq49" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.173392 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtwqv\" (UniqueName: \"kubernetes.io/projected/2190c8e2-2c7d-47fd-bb13-32ff0020c655-kube-api-access-jtwqv\") pod \"console-operator-58897d9998-7twx2\" (UID: \"2190c8e2-2c7d-47fd-bb13-32ff0020c655\") " pod="openshift-console-operator/console-operator-58897d9998-7twx2" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.194333 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpnkc\" (UniqueName: \"kubernetes.io/projected/138d7b9f-9cae-436f-a06e-e3ccd0b9168d-kube-api-access-lpnkc\") pod \"control-plane-machine-set-operator-78cbb6b69f-vjbjr\" (UID: \"138d7b9f-9cae-436f-a06e-e3ccd0b9168d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vjbjr" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.208582 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:06 crc kubenswrapper[4702]: E1125 13:48:06.208977 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:06.70895852 +0000 UTC m=+95.871608970 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.211772 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d41be55e-0563-404d-b2b0-d642b675910e-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-5vx95\" (UID: \"d41be55e-0563-404d-b2b0-d642b675910e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5vx95" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.222832 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xszrd" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.229143 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxhtc\" (UniqueName: \"kubernetes.io/projected/30dd0e87-b5d2-4ef3-896f-aaaaf64300ad-kube-api-access-jxhtc\") pod \"dns-default-jqmtw\" (UID: \"30dd0e87-b5d2-4ef3-896f-aaaaf64300ad\") " pod="openshift-dns/dns-default-jqmtw" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.230650 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mb6kd" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.245541 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8vr2l" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.249738 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mt6dk\" (UniqueName: \"kubernetes.io/projected/7de9c8c4-69ff-478f-9a13-ea39ca9a0db5-kube-api-access-mt6dk\") pod \"service-ca-9c57cc56f-j6d5l\" (UID: \"7de9c8c4-69ff-478f-9a13-ea39ca9a0db5\") " pod="openshift-service-ca/service-ca-9c57cc56f-j6d5l" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.261548 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gj88f" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.266654 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4zttp"] Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.278568 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/24dafe4a-2dce-4e4a-8a2c-d5653b326861-bound-sa-token\") pod \"ingress-operator-5b745b69d9-mlxwh\" (UID: \"24dafe4a-2dce-4e4a-8a2c-d5653b326861\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mlxwh" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.280301 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-mbmx2"] Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.283387 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-j6d5l" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.297549 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwrfb\" (UniqueName: \"kubernetes.io/projected/92fb9b54-8df1-40ac-8c8c-66d12416bc74-kube-api-access-hwrfb\") pod \"package-server-manager-789f6589d5-4zsfb\" (UID: \"92fb9b54-8df1-40ac-8c8c-66d12416bc74\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4zsfb" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.306259 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401305-j5rqf" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.312273 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:06 crc kubenswrapper[4702]: E1125 13:48:06.312786 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:06.812770514 +0000 UTC m=+95.975420964 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.313123 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vz97\" (UniqueName: \"kubernetes.io/projected/f51cb34e-c816-48dd-98d9-63ccc57222d7-kube-api-access-7vz97\") pod \"migrator-59844c95c7-twgvd\" (UID: \"f51cb34e-c816-48dd-98d9-63ccc57222d7\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-twgvd" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.314905 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-9bq59"] Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.320133 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wdn7d"] Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.321707 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mlxwh" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.327857 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-zqrgn" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.335411 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b2jt8"] Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.336459 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lkx85" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.336656 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvcqp\" (UniqueName: \"kubernetes.io/projected/17279013-fed3-475f-b960-3b9e41116da8-kube-api-access-dvcqp\") pod \"packageserver-d55dfcdfc-t5vtr\" (UID: \"17279013-fed3-475f-b960-3b9e41116da8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t5vtr" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.349800 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s"] Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.350005 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-lmq49" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.353335 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qk8s8\" (UniqueName: \"kubernetes.io/projected/9b430b6f-d294-4e5a-9410-fcdc8fca501d-kube-api-access-qk8s8\") pod \"multus-admission-controller-857f4d67dd-bmb9v\" (UID: \"9b430b6f-d294-4e5a-9410-fcdc8fca501d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-bmb9v" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.363854 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4zsfb" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.376188 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5829q\" (UniqueName: \"kubernetes.io/projected/a70d6c59-6b85-4950-8e30-1938da7a01e8-kube-api-access-5829q\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.380653 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b2jt8" event={"ID":"25f66de5-fdbd-44dc-911d-976d5c2bbfd3","Type":"ContainerStarted","Data":"3cd80492ddd3243b05d3a6da3fb867d102b107184d94264549850412183cd658"} Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.382917 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mbmx2" event={"ID":"dd97957f-e7df-48c2-b23f-b1101b786766","Type":"ContainerStarted","Data":"aa12f65cdcaa286d0c7dfbd14f4799d7d4dd526d3b29c1a9fe25334891dbc92c"} Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.383927 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-9bq59" event={"ID":"26e29bc4-82c6-47c1-8ffa-5e35e4b1c595","Type":"ContainerStarted","Data":"18e60dfba6ac942eca488476e42dff979b52da8734b433332c05050d7b86d648"} Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.386281 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-q2bqm" event={"ID":"5add632c-b0b2-490d-a0f6-d0193c176134","Type":"ContainerStarted","Data":"cbfb3fdc45452e17f1667ee22afd06143a554efa0b11070c2662a4b08f657797"} Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.386999 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wv6k6" event={"ID":"ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73","Type":"ContainerStarted","Data":"cdf2dd1f3ca165b58b9a99dbc1f7ad5e068a3d14b16ab3d05029a3402a409f4b"} Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.387778 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9xwkx" event={"ID":"e4c1d0e3-c3c8-47f3-ba79-d8aa252b3572","Type":"ContainerStarted","Data":"9fc51eee02caed6fce76cbc7e47be0f5a67c8ca26f3ae7e50af0ba64c4c70f60"} Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.390155 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" event={"ID":"ab12e899-20e3-4755-a537-3e54c5761af6","Type":"ContainerStarted","Data":"19775841748b298f15c092f075ddca7b6a703cd7a5d1f4487b5c1408c0878b4c"} Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.399935 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sxxzq" event={"ID":"3ce01e2e-eb84-47e0-a3c9-d4b0044f79a5","Type":"ContainerStarted","Data":"8aeb69ffdf0a23f5d3a69449c75b637bc4b9e5bb53be196c2b71e38a7c614f2f"} Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.399989 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sxxzq" event={"ID":"3ce01e2e-eb84-47e0-a3c9-d4b0044f79a5","Type":"ContainerStarted","Data":"7d1cf9991437ff417daec62e6ca82a48151d68513b2d68ea1154e2facf412f63"} Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.400565 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a70d6c59-6b85-4950-8e30-1938da7a01e8-bound-sa-token\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:06 crc kubenswrapper[4702]: W1125 13:48:06.408922 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb619039f_82c2_49ca_9319_75d0883a5a7d.slice/crio-43ce5e8755fd8c73cf3a1972da2f8e2bdd501a6db4d983596a2ead6fdd0cf029 WatchSource:0}: Error finding container 43ce5e8755fd8c73cf3a1972da2f8e2bdd501a6db4d983596a2ead6fdd0cf029: Status 404 returned error can't find the container with id 43ce5e8755fd8c73cf3a1972da2f8e2bdd501a6db4d983596a2ead6fdd0cf029 Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.411279 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-jqmtw" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.412878 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:06 crc kubenswrapper[4702]: E1125 13:48:06.413277 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:06.913221696 +0000 UTC m=+96.075872146 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.416708 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgr9k\" (UniqueName: \"kubernetes.io/projected/f8caa35c-4401-482b-8c5e-2927fbb50c4d-kube-api-access-vgr9k\") pod \"olm-operator-6b444d44fb-dk89m\" (UID: \"f8caa35c-4401-482b-8c5e-2927fbb50c4d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dk89m" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.421725 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-dnwh7"] Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.423452 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-pvqd2"] Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.460288 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rmtvx"] Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.461202 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hn9hb\" (UniqueName: \"kubernetes.io/projected/f961361a-3a5a-49be-b34a-06857d960b39-kube-api-access-hn9hb\") pod \"machine-config-server-8vt69\" (UID: \"f961361a-3a5a-49be-b34a-06857d960b39\") " pod="openshift-machine-config-operator/machine-config-server-8vt69" Nov 25 13:48:06 crc kubenswrapper[4702]: W1125 13:48:06.464391 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb22f225_d744_4357_a091_d21d76a31e42.slice/crio-e84f7b32813498f1145c64e5a192491686a0707b7f8ca88109db9d4764e2850b WatchSource:0}: Error finding container e84f7b32813498f1145c64e5a192491686a0707b7f8ca88109db9d4764e2850b: Status 404 returned error can't find the container with id e84f7b32813498f1145c64e5a192491686a0707b7f8ca88109db9d4764e2850b Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.468287 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bz2p7\" (UniqueName: \"kubernetes.io/projected/90b61112-2c79-4b92-930f-a182a569c0ae-kube-api-access-bz2p7\") pod \"ingress-canary-dkk9l\" (UID: \"90b61112-2c79-4b92-930f-a182a569c0ae\") " pod="openshift-ingress-canary/ingress-canary-dkk9l" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.469412 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-5tvj9"] Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.472121 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-7twx2" Nov 25 13:48:06 crc kubenswrapper[4702]: W1125 13:48:06.474452 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b821731_b65f_45df_b510_de915e573be1.slice/crio-bef57ad5ccf77cfcbc87f3637b12387f6ba3155b71234e140629494ddc3fafc5 WatchSource:0}: Error finding container bef57ad5ccf77cfcbc87f3637b12387f6ba3155b71234e140629494ddc3fafc5: Status 404 returned error can't find the container with id bef57ad5ccf77cfcbc87f3637b12387f6ba3155b71234e140629494ddc3fafc5 Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.491001 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vjbjr" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.505622 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5vx95" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.510780 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-twgvd" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.517722 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:06 crc kubenswrapper[4702]: E1125 13:48:06.518708 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:07.018214799 +0000 UTC m=+96.180865249 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.529948 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-bjl94"] Nov 25 13:48:06 crc kubenswrapper[4702]: W1125 13:48:06.546456 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a0fafb6_8df4_457c_9651_1d40cc9742b3.slice/crio-f70787c02ebbe5c9a65fdeb361e5850cba507867e58262632b5f537d7d284963 WatchSource:0}: Error finding container f70787c02ebbe5c9a65fdeb361e5850cba507867e58262632b5f537d7d284963: Status 404 returned error can't find the container with id f70787c02ebbe5c9a65fdeb361e5850cba507867e58262632b5f537d7d284963 Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.548368 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-c2xs7"] Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.572926 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-bmb9v" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.578497 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-5cmt4"] Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.615353 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-8j9xk"] Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.619342 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t5vtr" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.623538 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:06 crc kubenswrapper[4702]: E1125 13:48:06.624348 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:07.12433084 +0000 UTC m=+96.286981290 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.654048 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-j7l2t"] Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.680598 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dk89m" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.684421 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-8vr2l"] Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.722066 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-8vt69" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.722273 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xszrd"] Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.725211 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:06 crc kubenswrapper[4702]: E1125 13:48:06.725619 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:07.225599502 +0000 UTC m=+96.388250012 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.730938 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-dkk9l" Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.829109 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:06 crc kubenswrapper[4702]: E1125 13:48:06.829451 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:07.329405735 +0000 UTC m=+96.492056185 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.831268 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:06 crc kubenswrapper[4702]: E1125 13:48:06.832069 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:07.332050851 +0000 UTC m=+96.494701301 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:06 crc kubenswrapper[4702]: W1125 13:48:06.841261 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c290048_0c68_42fe_bd32_985795e837eb.slice/crio-cc6352c92a40c75dd785a6e5dd2586aec4489eef383f2dbac066babd9177a3f1 WatchSource:0}: Error finding container cc6352c92a40c75dd785a6e5dd2586aec4489eef383f2dbac066babd9177a3f1: Status 404 returned error can't find the container with id cc6352c92a40c75dd785a6e5dd2586aec4489eef383f2dbac066babd9177a3f1 Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.933639 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:06 crc kubenswrapper[4702]: E1125 13:48:06.935004 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:07.434972533 +0000 UTC m=+96.597622983 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.941981 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:06 crc kubenswrapper[4702]: E1125 13:48:06.943129 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:07.443107893 +0000 UTC m=+96.605758343 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.962353 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gj88f"] Nov 25 13:48:06 crc kubenswrapper[4702]: I1125 13:48:06.987657 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mb6kd"] Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.047128 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:07 crc kubenswrapper[4702]: E1125 13:48:07.047968 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:07.547946332 +0000 UTC m=+96.710596782 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.084277 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401305-j5rqf"] Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.084361 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-mlxwh"] Nov 25 13:48:07 crc kubenswrapper[4702]: W1125 13:48:07.134632 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcbf919e2_9e7d_4e22_a04d_e838be932151.slice/crio-30a036f631c8ebb7d80cb551bd85e6acdce4f4e88bb024da1a95bb554d9c3b19 WatchSource:0}: Error finding container 30a036f631c8ebb7d80cb551bd85e6acdce4f4e88bb024da1a95bb554d9c3b19: Status 404 returned error can't find the container with id 30a036f631c8ebb7d80cb551bd85e6acdce4f4e88bb024da1a95bb554d9c3b19 Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.149272 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:07 crc kubenswrapper[4702]: E1125 13:48:07.149629 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:07.649617404 +0000 UTC m=+96.812267854 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.194980 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-j6d5l"] Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.252186 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:07 crc kubenswrapper[4702]: E1125 13:48:07.252506 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:07.752490005 +0000 UTC m=+96.915140445 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.270204 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-zqrgn"] Nov 25 13:48:07 crc kubenswrapper[4702]: W1125 13:48:07.302525 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8cb1a228_426f_41cc_a15f_01dfb84a35cd.slice/crio-78e85639211e12a8fdb3314315d00af8eb3d9b5c47eb1338468bb8d99c4842d8 WatchSource:0}: Error finding container 78e85639211e12a8fdb3314315d00af8eb3d9b5c47eb1338468bb8d99c4842d8: Status 404 returned error can't find the container with id 78e85639211e12a8fdb3314315d00af8eb3d9b5c47eb1338468bb8d99c4842d8 Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.353544 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4zsfb"] Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.354764 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:07 crc kubenswrapper[4702]: E1125 13:48:07.355086 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:07.855071209 +0000 UTC m=+97.017721659 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.373688 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-twgvd"] Nov 25 13:48:07 crc kubenswrapper[4702]: W1125 13:48:07.421351 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0389edfa_9a86_48df_a9a6_26e2b4d092c4.slice/crio-1b59567bdfc4086d232bd13a1c3b486b274c4e0516721c3a81d9c98ca22ad672 WatchSource:0}: Error finding container 1b59567bdfc4086d232bd13a1c3b486b274c4e0516721c3a81d9c98ca22ad672: Status 404 returned error can't find the container with id 1b59567bdfc4086d232bd13a1c3b486b274c4e0516721c3a81d9c98ca22ad672 Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.427737 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b2jt8" event={"ID":"25f66de5-fdbd-44dc-911d-976d5c2bbfd3","Type":"ContainerStarted","Data":"42300817c336289b39543b31ac27e2f31d277131ede3403804c49aac875c3181"} Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.428126 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-jqmtw"] Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.430031 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-lmq49" event={"ID":"6d7fc28a-71f2-415c-a99e-76e4ab7fbb70","Type":"ContainerStarted","Data":"2d8e362fb843f70e8c9aa87760a368bf1a8b5a681abaf2551e4bc11d8e17a20f"} Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.434145 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lkx85"] Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.434979 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-5tvj9" event={"ID":"7a0fafb6-8df4-457c-9651-1d40cc9742b3","Type":"ContainerStarted","Data":"f70787c02ebbe5c9a65fdeb361e5850cba507867e58262632b5f537d7d284963"} Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.455883 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5vx95"] Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.456354 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:07 crc kubenswrapper[4702]: E1125 13:48:07.457539 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:07.95752453 +0000 UTC m=+97.120174980 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.462990 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wdn7d" event={"ID":"bb22f225-d744-4357-a091-d21d76a31e42","Type":"ContainerStarted","Data":"e84f7b32813498f1145c64e5a192491686a0707b7f8ca88109db9d4764e2850b"} Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.468987 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xszrd" event={"ID":"0c290048-0c68-42fe-bd32-985795e837eb","Type":"ContainerStarted","Data":"cc6352c92a40c75dd785a6e5dd2586aec4489eef383f2dbac066babd9177a3f1"} Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.471054 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gj88f" event={"ID":"89437c7c-7034-44f8-bb77-000a12163744","Type":"ContainerStarted","Data":"50b9b1bcb2f09f9473f7e51c691cbb22e5f3a3ee78f0b1ad7e1a1435c193825a"} Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.472177 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-8j9xk" event={"ID":"ce097f6c-2894-4417-b66b-0f30d91f3889","Type":"ContainerStarted","Data":"1002d2a82e51075d384ecded3e910e2110d5eb33884bf7c218c1172c44badb95"} Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.473514 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-dnwh7" event={"ID":"7cfd6b45-6486-4b59-9a79-21773426f02b","Type":"ContainerStarted","Data":"eb920a5439eb1b972625ce9c9da6b5a9dac5f07d91b3c8f60db75bf6c6dd5b5b"} Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.475945 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mlxwh" event={"ID":"24dafe4a-2dce-4e4a-8a2c-d5653b326861","Type":"ContainerStarted","Data":"7040b841a74e2c0c60620002ee2c24a2cd25d452afaf572f924ba43c7efe73fc"} Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.482910 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4zttp" event={"ID":"017380dd-ff61-49ce-8603-00daaeef7d5f","Type":"ContainerStarted","Data":"8e373521750b99a66b90751f4e6609e5eb148c2cfcd98242cbdc4ba01473d6e4"} Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.488074 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" event={"ID":"b619039f-82c2-49ca-9319-75d0883a5a7d","Type":"ContainerStarted","Data":"43ce5e8755fd8c73cf3a1972da2f8e2bdd501a6db4d983596a2ead6fdd0cf029"} Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.490898 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9xwkx" event={"ID":"e4c1d0e3-c3c8-47f3-ba79-d8aa252b3572","Type":"ContainerStarted","Data":"d955550b8809e67548a38625d36f99df3e7569d08a13184e17ab11bbc502cc34"} Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.496270 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-j7l2t" event={"ID":"f74394e4-d2c6-4211-be82-2f1c1de9c536","Type":"ContainerStarted","Data":"99223f5ff1625efa8ecd404e95779db535a69f064201002c0092613e943bb4f8"} Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.498777 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-q2bqm" event={"ID":"5add632c-b0b2-490d-a0f6-d0193c176134","Type":"ContainerStarted","Data":"bac7435e74a4afa3032853c97ce0c7d7490687e3df3be1b4f12581e00186e349"} Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.501839 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mb6kd" event={"ID":"cbf919e2-9e7d-4e22-a04d-e838be932151","Type":"ContainerStarted","Data":"30a036f631c8ebb7d80cb551bd85e6acdce4f4e88bb024da1a95bb554d9c3b19"} Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.503409 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wv6k6" event={"ID":"ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73","Type":"ContainerStarted","Data":"6afeef1a8b74e4384f98678148072a6d7b5248820dda22d75b4a48b81b6fd22f"} Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.504307 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wv6k6" Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.506248 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" event={"ID":"ab12e899-20e3-4755-a537-3e54c5761af6","Type":"ContainerStarted","Data":"cad53363793354aee615f19f39ca2f7c28c1b98cc6c7bbffd094efc5dab27bd5"} Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.506680 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.508744 4702 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-wv6k6 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.509153 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wv6k6" podUID="ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.508814 4702 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-dfr9q container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.34:6443/healthz\": dial tcp 10.217.0.34:6443: connect: connection refused" start-of-body= Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.509281 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" podUID="ab12e899-20e3-4755-a537-3e54c5761af6" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.34:6443/healthz\": dial tcp 10.217.0.34:6443: connect: connection refused" Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.513071 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401305-j5rqf" event={"ID":"8cb1a228-426f-41cc-a15f-01dfb84a35cd","Type":"ContainerStarted","Data":"78e85639211e12a8fdb3314315d00af8eb3d9b5c47eb1338468bb8d99c4842d8"} Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.516868 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" event={"ID":"8b821731-b65f-45df-b510-de915e573be1","Type":"ContainerStarted","Data":"bef57ad5ccf77cfcbc87f3637b12387f6ba3155b71234e140629494ddc3fafc5"} Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.518860 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-j6d5l" event={"ID":"7de9c8c4-69ff-478f-9a13-ea39ca9a0db5","Type":"ContainerStarted","Data":"e3d7763d83adc1ab816a7458afee33ee57ce4b86f186a9e8b5d4cac3edc381a3"} Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.528008 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8vr2l" event={"ID":"edba4243-475a-49b8-8d1d-183164f35a54","Type":"ContainerStarted","Data":"41ea2ab370a203cb115685efd40a21ba521143d870f2a65cac183cac8c7f71fb"} Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.532706 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bjl94" event={"ID":"d9232dc2-d71f-41b1-864a-82603860af69","Type":"ContainerStarted","Data":"d7efb521dad2719be25a16b47c422d72b8edd43121165be9e561589c1e671bbe"} Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.545022 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-c2xs7" event={"ID":"358a66f5-3192-4b20-ba92-37c8a8544209","Type":"ContainerStarted","Data":"3e8b249cbfafd3b10e85ffb47754a91b16c1337ff70ba21ca272b534b003f496"} Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.549741 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-9bq59" event={"ID":"26e29bc4-82c6-47c1-8ffa-5e35e4b1c595","Type":"ContainerStarted","Data":"85c2398dbeeb8b265a81a60844e5efc18da533fce022b08c573499ca5ca53329"} Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.553782 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mbmx2" event={"ID":"dd97957f-e7df-48c2-b23f-b1101b786766","Type":"ContainerStarted","Data":"378490d0ff8f1e48c7137212533d837ebb64ade1ba2b9ac710c6d34364c4bcf7"} Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.556704 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-5cmt4" event={"ID":"8f55435c-3542-4faf-997d-8a7dad39ce50","Type":"ContainerStarted","Data":"03d110a0677c2f84276e3ab2b2318f2e9dadcd0c4c946cb00e3c9d953862ea6c"} Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.557280 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.557563 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rmtvx" event={"ID":"5413dfbd-3c8b-42ac-8536-c0331fb7a397","Type":"ContainerStarted","Data":"a78343cf17571cc0a3b9d85ea6f4740c2bde3132db48e84cd73f03f40258c4dd"} Nov 25 13:48:07 crc kubenswrapper[4702]: E1125 13:48:07.558521 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:08.058506774 +0000 UTC m=+97.221157314 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.601764 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t5vtr"] Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.603732 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-7twx2"] Nov 25 13:48:07 crc kubenswrapper[4702]: W1125 13:48:07.647560 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod30dd0e87_b5d2_4ef3_896f_aaaaf64300ad.slice/crio-f84aafcbd07e595f9bafb2b36bb3e99715c4569ffe78ac72a3ee824c659158bf WatchSource:0}: Error finding container f84aafcbd07e595f9bafb2b36bb3e99715c4569ffe78ac72a3ee824c659158bf: Status 404 returned error can't find the container with id f84aafcbd07e595f9bafb2b36bb3e99715c4569ffe78ac72a3ee824c659158bf Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.671041 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:07 crc kubenswrapper[4702]: E1125 13:48:07.672023 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:08.171999267 +0000 UTC m=+97.334649717 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.678185 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:07 crc kubenswrapper[4702]: E1125 13:48:07.679696 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:08.179681446 +0000 UTC m=+97.342331896 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.704175 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vjbjr"] Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.715744 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dk89m"] Nov 25 13:48:07 crc kubenswrapper[4702]: W1125 13:48:07.716246 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2190c8e2_2c7d_47fd_bb13_32ff0020c655.slice/crio-64b2c0ebce5797a75bd493bd2298dfef3c0110a039c3f0b82353641342026c68 WatchSource:0}: Error finding container 64b2c0ebce5797a75bd493bd2298dfef3c0110a039c3f0b82353641342026c68: Status 404 returned error can't find the container with id 64b2c0ebce5797a75bd493bd2298dfef3c0110a039c3f0b82353641342026c68 Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.757103 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-bmb9v"] Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.773833 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-dkk9l"] Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.780640 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:07 crc kubenswrapper[4702]: E1125 13:48:07.781403 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:08.281374537 +0000 UTC m=+97.444024987 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.781698 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:07 crc kubenswrapper[4702]: E1125 13:48:07.782138 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:08.282126456 +0000 UTC m=+97.444776906 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:07 crc kubenswrapper[4702]: W1125 13:48:07.809651 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b430b6f_d294_4e5a_9410_fcdc8fca501d.slice/crio-b5e5619425e760b95e711d30961501058a65335345dd7299781da4dc314d16a7 WatchSource:0}: Error finding container b5e5619425e760b95e711d30961501058a65335345dd7299781da4dc314d16a7: Status 404 returned error can't find the container with id b5e5619425e760b95e711d30961501058a65335345dd7299781da4dc314d16a7 Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.882742 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:07 crc kubenswrapper[4702]: E1125 13:48:07.882917 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:08.382887416 +0000 UTC m=+97.545537866 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.883178 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:07 crc kubenswrapper[4702]: E1125 13:48:07.883699 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:08.383683685 +0000 UTC m=+97.546334125 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.889472 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wv6k6" podStartSLOduration=71.889455226 podStartE2EDuration="1m11.889455226s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:07.888662578 +0000 UTC m=+97.051313048" watchObservedRunningTime="2025-11-25 13:48:07.889455226 +0000 UTC m=+97.052105676" Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.936029 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-mbmx2" podStartSLOduration=71.936007823 podStartE2EDuration="1m11.936007823s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:07.932578248 +0000 UTC m=+97.095228698" watchObservedRunningTime="2025-11-25 13:48:07.936007823 +0000 UTC m=+97.098658273" Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.971774 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b2jt8" podStartSLOduration=71.971738231 podStartE2EDuration="1m11.971738231s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:07.971567997 +0000 UTC m=+97.134218447" watchObservedRunningTime="2025-11-25 13:48:07.971738231 +0000 UTC m=+97.134388691" Nov 25 13:48:07 crc kubenswrapper[4702]: I1125 13:48:07.985998 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:07 crc kubenswrapper[4702]: E1125 13:48:07.986567 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:08.486547496 +0000 UTC m=+97.649197936 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.013708 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" podStartSLOduration=73.013691063 podStartE2EDuration="1m13.013691063s" podCreationTimestamp="2025-11-25 13:46:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:08.012066934 +0000 UTC m=+97.174717404" watchObservedRunningTime="2025-11-25 13:48:08.013691063 +0000 UTC m=+97.176341513" Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.093030 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:08 crc kubenswrapper[4702]: E1125 13:48:08.093439 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:08.593425605 +0000 UTC m=+97.756076055 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.194544 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:08 crc kubenswrapper[4702]: E1125 13:48:08.194704 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:08.694680806 +0000 UTC m=+97.857331256 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.195125 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:08 crc kubenswrapper[4702]: E1125 13:48:08.195466 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:08.695457916 +0000 UTC m=+97.858108366 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.301377 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:08 crc kubenswrapper[4702]: E1125 13:48:08.302259 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:08.802215752 +0000 UTC m=+97.964866202 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.404861 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:08 crc kubenswrapper[4702]: E1125 13:48:08.410015 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:08.909988254 +0000 UTC m=+98.072638704 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.505799 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:08 crc kubenswrapper[4702]: E1125 13:48:08.506009 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:09.005981426 +0000 UTC m=+98.168631876 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.506419 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:08 crc kubenswrapper[4702]: E1125 13:48:08.506867 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:09.006843347 +0000 UTC m=+98.169493787 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.604919 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-8j9xk" event={"ID":"ce097f6c-2894-4417-b66b-0f30d91f3889","Type":"ContainerStarted","Data":"24f59c64cc8ef79a1d46982ef248ab20ca74509e0a6c54165c7f37b355b0bc3b"} Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.608224 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-8j9xk" Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.610367 4702 patch_prober.go:28] interesting pod/downloads-7954f5f757-8j9xk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.610425 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8j9xk" podUID="ce097f6c-2894-4417-b66b-0f30d91f3889" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.614533 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:08 crc kubenswrapper[4702]: E1125 13:48:08.615094 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:09.11505702 +0000 UTC m=+98.277707470 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.618821 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lkx85" event={"ID":"ec7a64ba-34c9-4e31-8e2b-1991cba108d0","Type":"ContainerStarted","Data":"1cf15146394cd2088c969d2684b4f07a566db17eda7508bd54d0b7b8353b45ae"} Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.623190 4702 generic.go:334] "Generic (PLEG): container finished" podID="8b821731-b65f-45df-b510-de915e573be1" containerID="b03e8a6c3a0b8b54311016a628af9d8eb560349f71f43fca1d2a8d568060e31f" exitCode=0 Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.623287 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" event={"ID":"8b821731-b65f-45df-b510-de915e573be1","Type":"ContainerDied","Data":"b03e8a6c3a0b8b54311016a628af9d8eb560349f71f43fca1d2a8d568060e31f"} Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.630489 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-8j9xk" podStartSLOduration=72.630461739 podStartE2EDuration="1m12.630461739s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:08.629904424 +0000 UTC m=+97.792554874" watchObservedRunningTime="2025-11-25 13:48:08.630461739 +0000 UTC m=+97.793112199" Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.630788 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9xwkx" podStartSLOduration=72.630781616 podStartE2EDuration="1m12.630781616s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:08.055495392 +0000 UTC m=+97.218145842" watchObservedRunningTime="2025-11-25 13:48:08.630781616 +0000 UTC m=+97.793432066" Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.647972 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-jqmtw" event={"ID":"30dd0e87-b5d2-4ef3-896f-aaaaf64300ad","Type":"ContainerStarted","Data":"f84aafcbd07e595f9bafb2b36bb3e99715c4569ffe78ac72a3ee824c659158bf"} Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.669390 4702 generic.go:334] "Generic (PLEG): container finished" podID="358a66f5-3192-4b20-ba92-37c8a8544209" containerID="fae532ca1a43271fcd1057133b9acbd9c6587bd5c95178eb016b3ee320605d57" exitCode=0 Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.669527 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-c2xs7" event={"ID":"358a66f5-3192-4b20-ba92-37c8a8544209","Type":"ContainerDied","Data":"fae532ca1a43271fcd1057133b9acbd9c6587bd5c95178eb016b3ee320605d57"} Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.683682 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-lmq49" event={"ID":"6d7fc28a-71f2-415c-a99e-76e4ab7fbb70","Type":"ContainerStarted","Data":"71b00924a7977d899aaf619ad16fb0bed84fd5af4aa212af10c97d8727004d08"} Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.704257 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rmtvx" event={"ID":"5413dfbd-3c8b-42ac-8536-c0331fb7a397","Type":"ContainerStarted","Data":"d5cab50337cc399fb7170add9ede78cdd42dea6d9c62d3d56a2c545af1c532e4"} Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.704515 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-rmtvx" Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.719811 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.720333 4702 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-rmtvx container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.720403 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-rmtvx" podUID="5413dfbd-3c8b-42ac-8536-c0331fb7a397" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Nov 25 13:48:08 crc kubenswrapper[4702]: E1125 13:48:08.721452 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:09.221436397 +0000 UTC m=+98.384086847 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.738694 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-twgvd" event={"ID":"f51cb34e-c816-48dd-98d9-63ccc57222d7","Type":"ContainerStarted","Data":"167adf51c4301148aaf82155008673409d0f96adec45d7ca3192c51857b61462"} Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.738752 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-twgvd" event={"ID":"f51cb34e-c816-48dd-98d9-63ccc57222d7","Type":"ContainerStarted","Data":"05ec057db2f838f863ba94b00cd29363091993d49119934202c39a40f3920098"} Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.747866 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-8vt69" event={"ID":"f961361a-3a5a-49be-b34a-06857d960b39","Type":"ContainerStarted","Data":"594396e116971c020b100f60008ca5d0790df452eb4de3652c70b1a421a5bc60"} Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.753914 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-dnwh7" event={"ID":"7cfd6b45-6486-4b59-9a79-21773426f02b","Type":"ContainerStarted","Data":"167e898c937c91879550b12669580f04ee93523f2a93bff2c6bac225df1a5596"} Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.764052 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xszrd" event={"ID":"0c290048-0c68-42fe-bd32-985795e837eb","Type":"ContainerStarted","Data":"e76acb4f952af33d9a105dbd80b9989dc67b4c0e36764571ce748ed7aa3fb667"} Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.765672 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-zqrgn" event={"ID":"0389edfa-9a86-48df-a9a6-26e2b4d092c4","Type":"ContainerStarted","Data":"1b59567bdfc4086d232bd13a1c3b486b274c4e0516721c3a81d9c98ca22ad672"} Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.775668 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dk89m" event={"ID":"f8caa35c-4401-482b-8c5e-2927fbb50c4d","Type":"ContainerStarted","Data":"67d7cc84c70a16f1d1864d9284ffc8c30a5b38500f229b598ef7ff1e133305c1"} Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.787761 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-lmq49" podStartSLOduration=72.787735718 podStartE2EDuration="1m12.787735718s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:08.77970292 +0000 UTC m=+97.942353390" watchObservedRunningTime="2025-11-25 13:48:08.787735718 +0000 UTC m=+97.950386168" Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.788435 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-rmtvx" podStartSLOduration=72.788428554 podStartE2EDuration="1m12.788428554s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:08.74269044 +0000 UTC m=+97.905340890" watchObservedRunningTime="2025-11-25 13:48:08.788428554 +0000 UTC m=+97.951079004" Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.795289 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bjl94" event={"ID":"d9232dc2-d71f-41b1-864a-82603860af69","Type":"ContainerStarted","Data":"3b5d0299b14e119ad0d65e8e7fd26aabb1593e3e5ef8fb5b6cdaa738ff763ada"} Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.820002 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-7twx2" event={"ID":"2190c8e2-2c7d-47fd-bb13-32ff0020c655","Type":"ContainerStarted","Data":"64b2c0ebce5797a75bd493bd2298dfef3c0110a039c3f0b82353641342026c68"} Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.820214 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xszrd" podStartSLOduration=72.820182916 podStartE2EDuration="1m12.820182916s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:08.812975679 +0000 UTC m=+97.975626129" watchObservedRunningTime="2025-11-25 13:48:08.820182916 +0000 UTC m=+97.982833366" Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.821081 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-7twx2" Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.822183 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.826607 4702 patch_prober.go:28] interesting pod/console-operator-58897d9998-7twx2 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/readyz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.826661 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-7twx2" podUID="2190c8e2-2c7d-47fd-bb13-32ff0020c655" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/readyz\": dial tcp 10.217.0.18:8443: connect: connection refused" Nov 25 13:48:08 crc kubenswrapper[4702]: E1125 13:48:08.834728 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:09.334665902 +0000 UTC m=+98.497316352 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.856986 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wdn7d" event={"ID":"bb22f225-d744-4357-a091-d21d76a31e42","Type":"ContainerStarted","Data":"d0e02825d6fa4d1bf22592aa82afbf04cc8939379d87d3e1b6ef54e1d1187d52"} Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.861542 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-dnwh7" podStartSLOduration=73.861516643 podStartE2EDuration="1m13.861516643s" podCreationTimestamp="2025-11-25 13:46:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:08.856492969 +0000 UTC m=+98.019143419" watchObservedRunningTime="2025-11-25 13:48:08.861516643 +0000 UTC m=+98.024167093" Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.881073 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t5vtr" event={"ID":"17279013-fed3-475f-b960-3b9e41116da8","Type":"ContainerStarted","Data":"1951b3e007cf0220844e2adc22c9d360be52649cb8dbf65ab03956d325940214"} Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.886770 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-7twx2" podStartSLOduration=72.886748754 podStartE2EDuration="1m12.886748754s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:08.885094803 +0000 UTC m=+98.047745273" watchObservedRunningTime="2025-11-25 13:48:08.886748754 +0000 UTC m=+98.049399204" Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.913224 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wdn7d" podStartSLOduration=73.913206935 podStartE2EDuration="1m13.913206935s" podCreationTimestamp="2025-11-25 13:46:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:08.911334949 +0000 UTC m=+98.073985399" watchObservedRunningTime="2025-11-25 13:48:08.913206935 +0000 UTC m=+98.075857385" Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.921956 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gj88f" event={"ID":"89437c7c-7034-44f8-bb77-000a12163744","Type":"ContainerStarted","Data":"c61793bffa69b9ecfe91f23b2cbe205a011941165181ae57b1ff104612dbffc9"} Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.923301 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gj88f" Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.925543 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:08 crc kubenswrapper[4702]: E1125 13:48:08.926575 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:09.426561503 +0000 UTC m=+98.589212033 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.938791 4702 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-gj88f container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.938851 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gj88f" podUID="89437c7c-7034-44f8-bb77-000a12163744" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.947563 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-5tvj9" event={"ID":"7a0fafb6-8df4-457c-9651-1d40cc9742b3","Type":"ContainerStarted","Data":"2d8b859abc57ea82cd4c273d2394a8cd1bfe9e7f7cd0f5168ec9ee93c1f38b62"} Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.974990 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gj88f" podStartSLOduration=72.974968225 podStartE2EDuration="1m12.974968225s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:08.9735467 +0000 UTC m=+98.136197170" watchObservedRunningTime="2025-11-25 13:48:08.974968225 +0000 UTC m=+98.137618675" Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.978720 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sxxzq" event={"ID":"3ce01e2e-eb84-47e0-a3c9-d4b0044f79a5","Type":"ContainerStarted","Data":"650e0b18b06103b6166718f4d11090689233c4a8d0f91faef1c158b6827cff6f"} Nov 25 13:48:08 crc kubenswrapper[4702]: I1125 13:48:08.985647 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5vx95" event={"ID":"d41be55e-0563-404d-b2b0-d642b675910e","Type":"ContainerStarted","Data":"8dad402e93a2d8369f32708cf5658ac2219ff23749b334ec477b3dad023f6c79"} Nov 25 13:48:09 crc kubenswrapper[4702]: I1125 13:48:09.004406 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vjbjr" event={"ID":"138d7b9f-9cae-436f-a06e-e3ccd0b9168d","Type":"ContainerStarted","Data":"d68c1cbf1069118f5b56ce23cc25ad4b4f237c4bb5d3a58401c70f3e47987faa"} Nov 25 13:48:09 crc kubenswrapper[4702]: I1125 13:48:09.005875 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-5tvj9" podStartSLOduration=73.005858245 podStartE2EDuration="1m13.005858245s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:09.004721097 +0000 UTC m=+98.167371547" watchObservedRunningTime="2025-11-25 13:48:09.005858245 +0000 UTC m=+98.168508695" Nov 25 13:48:09 crc kubenswrapper[4702]: I1125 13:48:09.026451 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:09 crc kubenswrapper[4702]: E1125 13:48:09.027767 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:09.527748663 +0000 UTC m=+98.690399103 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:09 crc kubenswrapper[4702]: I1125 13:48:09.029738 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4zttp" event={"ID":"017380dd-ff61-49ce-8603-00daaeef7d5f","Type":"ContainerStarted","Data":"110ee89df17c29e4336477ebf782fd2ceccf72cd13c867b1feb859184f9d0422"} Nov 25 13:48:09 crc kubenswrapper[4702]: I1125 13:48:09.073223 4702 generic.go:334] "Generic (PLEG): container finished" podID="b619039f-82c2-49ca-9319-75d0883a5a7d" containerID="ccb2173e3f4ab4f9b239b657a851fa0c94a4e189c227c9a369a36aeb92906a5d" exitCode=0 Nov 25 13:48:09 crc kubenswrapper[4702]: I1125 13:48:09.073513 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" event={"ID":"b619039f-82c2-49ca-9319-75d0883a5a7d","Type":"ContainerDied","Data":"ccb2173e3f4ab4f9b239b657a851fa0c94a4e189c227c9a369a36aeb92906a5d"} Nov 25 13:48:09 crc kubenswrapper[4702]: I1125 13:48:09.079615 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8vr2l" event={"ID":"edba4243-475a-49b8-8d1d-183164f35a54","Type":"ContainerStarted","Data":"6bf48196d0e69c5e7ef3e836d41973e67600bb94bec5c76c4a7ce10b4901c69d"} Nov 25 13:48:09 crc kubenswrapper[4702]: I1125 13:48:09.092451 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-dkk9l" event={"ID":"90b61112-2c79-4b92-930f-a182a569c0ae","Type":"ContainerStarted","Data":"dbdcb2e9680ff193cd030344c6cdc1d419b5e8bf7fc3edd36afd727a4b4c03c5"} Nov 25 13:48:09 crc kubenswrapper[4702]: I1125 13:48:09.096080 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sxxzq" podStartSLOduration=74.096034583 podStartE2EDuration="1m14.096034583s" podCreationTimestamp="2025-11-25 13:46:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:09.064054276 +0000 UTC m=+98.226704726" watchObservedRunningTime="2025-11-25 13:48:09.096034583 +0000 UTC m=+98.258685233" Nov 25 13:48:09 crc kubenswrapper[4702]: I1125 13:48:09.110638 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-bmb9v" event={"ID":"9b430b6f-d294-4e5a-9410-fcdc8fca501d","Type":"ContainerStarted","Data":"b5e5619425e760b95e711d30961501058a65335345dd7299781da4dc314d16a7"} Nov 25 13:48:09 crc kubenswrapper[4702]: I1125 13:48:09.119371 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4zsfb" event={"ID":"92fb9b54-8df1-40ac-8c8c-66d12416bc74","Type":"ContainerStarted","Data":"db585ae2bd3a8cbf61d8f10012111199ada2d5170ff906a7623b0cdb74dbe36b"} Nov 25 13:48:09 crc kubenswrapper[4702]: I1125 13:48:09.129727 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:09 crc kubenswrapper[4702]: E1125 13:48:09.131086 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:09.631074635 +0000 UTC m=+98.793725085 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:09 crc kubenswrapper[4702]: I1125 13:48:09.135743 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-dkk9l" podStartSLOduration=6.135727659 podStartE2EDuration="6.135727659s" podCreationTimestamp="2025-11-25 13:48:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:09.135105295 +0000 UTC m=+98.297755745" watchObservedRunningTime="2025-11-25 13:48:09.135727659 +0000 UTC m=+98.298378109" Nov 25 13:48:09 crc kubenswrapper[4702]: I1125 13:48:09.136357 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-j7l2t" event={"ID":"f74394e4-d2c6-4211-be82-2f1c1de9c536","Type":"ContainerStarted","Data":"95d8cc680d66c8da89b6abbbbee6fe293bb3e767057bc6fd36b099dde4fd6de0"} Nov 25 13:48:09 crc kubenswrapper[4702]: I1125 13:48:09.151419 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wv6k6" Nov 25 13:48:09 crc kubenswrapper[4702]: I1125 13:48:09.165500 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:48:09 crc kubenswrapper[4702]: I1125 13:48:09.204730 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-j7l2t" podStartSLOduration=73.204711167 podStartE2EDuration="1m13.204711167s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:09.173302824 +0000 UTC m=+98.335953274" watchObservedRunningTime="2025-11-25 13:48:09.204711167 +0000 UTC m=+98.367361617" Nov 25 13:48:09 crc kubenswrapper[4702]: I1125 13:48:09.231017 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:09 crc kubenswrapper[4702]: E1125 13:48:09.232445 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:09.732420909 +0000 UTC m=+98.895071359 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:09 crc kubenswrapper[4702]: I1125 13:48:09.335779 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:09 crc kubenswrapper[4702]: E1125 13:48:09.336613 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:09.836598592 +0000 UTC m=+98.999249042 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:09 crc kubenswrapper[4702]: I1125 13:48:09.353807 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-lmq49" Nov 25 13:48:09 crc kubenswrapper[4702]: I1125 13:48:09.364612 4702 patch_prober.go:28] interesting pod/router-default-5444994796-lmq49 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 13:48:09 crc kubenswrapper[4702]: [-]has-synced failed: reason withheld Nov 25 13:48:09 crc kubenswrapper[4702]: [+]process-running ok Nov 25 13:48:09 crc kubenswrapper[4702]: healthz check failed Nov 25 13:48:09 crc kubenswrapper[4702]: I1125 13:48:09.364656 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lmq49" podUID="6d7fc28a-71f2-415c-a99e-76e4ab7fbb70" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 13:48:09 crc kubenswrapper[4702]: I1125 13:48:09.437102 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:09 crc kubenswrapper[4702]: E1125 13:48:09.437373 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:09.937354431 +0000 UTC m=+99.100004891 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:09 crc kubenswrapper[4702]: I1125 13:48:09.437587 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:09 crc kubenswrapper[4702]: E1125 13:48:09.438030 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:09.938020147 +0000 UTC m=+99.100670597 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:09 crc kubenswrapper[4702]: I1125 13:48:09.539858 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:09 crc kubenswrapper[4702]: E1125 13:48:09.540103 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:10.040072538 +0000 UTC m=+99.202723008 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:09 crc kubenswrapper[4702]: I1125 13:48:09.540771 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:09 crc kubenswrapper[4702]: E1125 13:48:09.541208 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:10.041192156 +0000 UTC m=+99.203842606 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:09 crc kubenswrapper[4702]: I1125 13:48:09.643343 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:09 crc kubenswrapper[4702]: E1125 13:48:09.644052 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:10.144018325 +0000 UTC m=+99.306668775 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:09 crc kubenswrapper[4702]: I1125 13:48:09.745140 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:09 crc kubenswrapper[4702]: E1125 13:48:09.745823 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:10.24580835 +0000 UTC m=+99.408458800 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:09 crc kubenswrapper[4702]: I1125 13:48:09.850393 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:09 crc kubenswrapper[4702]: E1125 13:48:09.859116 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:10.349207284 +0000 UTC m=+99.511857734 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:09 crc kubenswrapper[4702]: I1125 13:48:09.859538 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:09 crc kubenswrapper[4702]: E1125 13:48:09.860166 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:10.360153364 +0000 UTC m=+99.522803814 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:09 crc kubenswrapper[4702]: I1125 13:48:09.964249 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:09 crc kubenswrapper[4702]: E1125 13:48:09.964802 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:10.464765868 +0000 UTC m=+99.627416318 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.066167 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:10 crc kubenswrapper[4702]: E1125 13:48:10.066848 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:10.566828488 +0000 UTC m=+99.729478938 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.143865 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4zttp" event={"ID":"017380dd-ff61-49ce-8603-00daaeef7d5f","Type":"ContainerStarted","Data":"b6750fa89e0f9f16d15c8c32f4138044b637c176ff9289262045f13099c3d3d5"} Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.147650 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-zqrgn" event={"ID":"0389edfa-9a86-48df-a9a6-26e2b4d092c4","Type":"ContainerStarted","Data":"1ddbe9a5a1c75085ea60f719d36e9a81327d12f38b34f6072fd550416ba6c8f8"} Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.148592 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-zqrgn" Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.150567 4702 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-zqrgn container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.150610 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-zqrgn" podUID="0389edfa-9a86-48df-a9a6-26e2b4d092c4" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.155080 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-q2bqm" event={"ID":"5add632c-b0b2-490d-a0f6-d0193c176134","Type":"ContainerStarted","Data":"2ed9e7c06595094ee1272f98094d4418873b1e576d6a95aa830a591bdb1f19cd"} Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.157759 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-twgvd" event={"ID":"f51cb34e-c816-48dd-98d9-63ccc57222d7","Type":"ContainerStarted","Data":"aeb15c761ce7bc6b0cca197204480afc52f0c12f4e425b028ff07f99b0a2d195"} Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.161191 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vjbjr" event={"ID":"138d7b9f-9cae-436f-a06e-e3ccd0b9168d","Type":"ContainerStarted","Data":"9123c49fade1885f7354cd0d042143f7d263af9974d192c7c1ebfe0065890807"} Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.162693 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401305-j5rqf" event={"ID":"8cb1a228-426f-41cc-a15f-01dfb84a35cd","Type":"ContainerStarted","Data":"c8fc47a36657d5c1b5bd76bdeca4e43395474e5ef3305ab606d2c7edfe968fa0"} Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.164172 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-jqmtw" event={"ID":"30dd0e87-b5d2-4ef3-896f-aaaaf64300ad","Type":"ContainerStarted","Data":"16ca8f5e26a3852403bf0c74f280ea2b6562fc017cd37e090ae8ba504557d489"} Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.164199 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-jqmtw" event={"ID":"30dd0e87-b5d2-4ef3-896f-aaaaf64300ad","Type":"ContainerStarted","Data":"44f61294c9804cd5a1b849ff2cf5c4a5a536a5444fe9e7c0c776b2273059e9fe"} Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.164576 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-jqmtw" Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.165790 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-dkk9l" event={"ID":"90b61112-2c79-4b92-930f-a182a569c0ae","Type":"ContainerStarted","Data":"3ca75b7cfb438e8253fac49843127d678f1e043c00e2fb9039a723e8790373e0"} Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.166721 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:10 crc kubenswrapper[4702]: E1125 13:48:10.167216 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:10.667197438 +0000 UTC m=+99.829847888 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.189616 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-c2xs7" event={"ID":"358a66f5-3192-4b20-ba92-37c8a8544209","Type":"ContainerStarted","Data":"f98ac44765b0d2df2226e151bd0a865a8044ba0f61dd50b09d2c986dff7410b9"} Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.189681 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-c2xs7" Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.210690 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4zttp" podStartSLOduration=74.210671727 podStartE2EDuration="1m14.210671727s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:10.173622256 +0000 UTC m=+99.336272706" watchObservedRunningTime="2025-11-25 13:48:10.210671727 +0000 UTC m=+99.373322177" Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.212081 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vjbjr" podStartSLOduration=74.212076762 podStartE2EDuration="1m14.212076762s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:10.20913671 +0000 UTC m=+99.371787160" watchObservedRunningTime="2025-11-25 13:48:10.212076762 +0000 UTC m=+99.374727212" Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.218383 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-9bq59" event={"ID":"26e29bc4-82c6-47c1-8ffa-5e35e4b1c595","Type":"ContainerStarted","Data":"e8c001ae58b49cec7ab739efc43c33048428740a415aeea8e38dac27840794f2"} Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.220906 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mlxwh" event={"ID":"24dafe4a-2dce-4e4a-8a2c-d5653b326861","Type":"ContainerStarted","Data":"319f063a4d7d95c4577b33d963bda810d9b8824f4603be57ffc5ea46da132f42"} Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.220961 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mlxwh" event={"ID":"24dafe4a-2dce-4e4a-8a2c-d5653b326861","Type":"ContainerStarted","Data":"a3767fb599b07abeb9e5f1052e456804307eff7686509875ae1260ce53f2adb0"} Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.229550 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lkx85" event={"ID":"ec7a64ba-34c9-4e31-8e2b-1991cba108d0","Type":"ContainerStarted","Data":"c1177aad8d08f439643e43e08062eefa88a5376b7aafa8aa5ed34b3745626b81"} Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.238592 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bjl94" event={"ID":"d9232dc2-d71f-41b1-864a-82603860af69","Type":"ContainerStarted","Data":"1c8675f64c898fee2eca99be77897f4ff72c5fc5cad1d083b69c5e675e3193d1"} Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.241304 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mb6kd" event={"ID":"cbf919e2-9e7d-4e22-a04d-e838be932151","Type":"ContainerStarted","Data":"9dbb55ce8a3feeeda7772b9b4e8a30c1da3b4ce2c5cb080c16c4420c44607aaf"} Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.243879 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4zsfb" event={"ID":"92fb9b54-8df1-40ac-8c8c-66d12416bc74","Type":"ContainerStarted","Data":"12a8aa8f4eadb2b21c7b0ad6f79d56edcde7e64a38363153d0a542d1e63bc251"} Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.243909 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4zsfb" event={"ID":"92fb9b54-8df1-40ac-8c8c-66d12416bc74","Type":"ContainerStarted","Data":"e7810cdfa1eca406d9660aacca9ff4a45eb6b6ecb1aa5fc836bf4541df6e6708"} Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.244350 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4zsfb" Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.269821 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:10 crc kubenswrapper[4702]: E1125 13:48:10.271426 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:10.771410092 +0000 UTC m=+99.934060542 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.271647 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-zqrgn" podStartSLOduration=74.271614977 podStartE2EDuration="1m14.271614977s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:10.252588089 +0000 UTC m=+99.415238539" watchObservedRunningTime="2025-11-25 13:48:10.271614977 +0000 UTC m=+99.434265417" Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.284477 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8vr2l" event={"ID":"edba4243-475a-49b8-8d1d-183164f35a54","Type":"ContainerStarted","Data":"fbed06e2120752bc177ef4d461af92b32480498410c93f310f8894b59417cb93"} Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.290895 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29401305-j5rqf" podStartSLOduration=75.29087483 podStartE2EDuration="1m15.29087483s" podCreationTimestamp="2025-11-25 13:46:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:10.289912977 +0000 UTC m=+99.452563417" watchObservedRunningTime="2025-11-25 13:48:10.29087483 +0000 UTC m=+99.453525280" Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.302173 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5vx95" event={"ID":"d41be55e-0563-404d-b2b0-d642b675910e","Type":"ContainerStarted","Data":"503ba1ed63b3550f8a138e55383615f4dbdfb149646ee6c871b657ae71244817"} Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.317578 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-7twx2" event={"ID":"2190c8e2-2c7d-47fd-bb13-32ff0020c655","Type":"ContainerStarted","Data":"88a9e379acb4c5c626aaadab08967ecbe5da8670fb5e1f1425c06fde7f6fc2b4"} Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.318700 4702 patch_prober.go:28] interesting pod/console-operator-58897d9998-7twx2 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/readyz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.318732 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-7twx2" podUID="2190c8e2-2c7d-47fd-bb13-32ff0020c655" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/readyz\": dial tcp 10.217.0.18:8443: connect: connection refused" Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.337461 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-8vt69" event={"ID":"f961361a-3a5a-49be-b34a-06857d960b39","Type":"ContainerStarted","Data":"66ee95efa38a76eb4140931bc4fd58a01d4cd0e6ad720ab632b8f4e720dc7fe8"} Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.349120 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-5cmt4" event={"ID":"8f55435c-3542-4faf-997d-8a7dad39ce50","Type":"ContainerStarted","Data":"52392b4be111d95eca73fa5e7befb2b4c9c1d018690bda52e48d62da99716850"} Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.351961 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-twgvd" podStartSLOduration=74.351944763 podStartE2EDuration="1m14.351944763s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:10.325270747 +0000 UTC m=+99.487921197" watchObservedRunningTime="2025-11-25 13:48:10.351944763 +0000 UTC m=+99.514595213" Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.356136 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" event={"ID":"b619039f-82c2-49ca-9319-75d0883a5a7d","Type":"ContainerStarted","Data":"6dbf8a4a08d8485637b6a003ceb09508d335a0eb480f102a3f75fea79489e460"} Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.357134 4702 patch_prober.go:28] interesting pod/router-default-5444994796-lmq49 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 13:48:10 crc kubenswrapper[4702]: [-]has-synced failed: reason withheld Nov 25 13:48:10 crc kubenswrapper[4702]: [+]process-running ok Nov 25 13:48:10 crc kubenswrapper[4702]: healthz check failed Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.357184 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lmq49" podUID="6d7fc28a-71f2-415c-a99e-76e4ab7fbb70" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.374591 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:10 crc kubenswrapper[4702]: E1125 13:48:10.377492 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:10.877472701 +0000 UTC m=+100.040123151 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.384652 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-q2bqm" podStartSLOduration=74.384627547 podStartE2EDuration="1m14.384627547s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:10.352766574 +0000 UTC m=+99.515417014" watchObservedRunningTime="2025-11-25 13:48:10.384627547 +0000 UTC m=+99.547277987" Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.392037 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" event={"ID":"8b821731-b65f-45df-b510-de915e573be1","Type":"ContainerStarted","Data":"d759d9cec575aa29423e717222d213a1572a58be26c91667135f2f4e5e143a57"} Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.405651 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-j6d5l" event={"ID":"7de9c8c4-69ff-478f-9a13-ea39ca9a0db5","Type":"ContainerStarted","Data":"595dda39fa00042df8735236525ff1b238e2267f4a558143638c1369dd4fe7aa"} Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.413803 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-bmb9v" event={"ID":"9b430b6f-d294-4e5a-9410-fcdc8fca501d","Type":"ContainerStarted","Data":"2a6327d89977dfe90e3a511ebb027de9a8a530adf7c45fa1b660b8d6d2377939"} Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.413851 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-bmb9v" event={"ID":"9b430b6f-d294-4e5a-9410-fcdc8fca501d","Type":"ContainerStarted","Data":"b8361ed06fae672cdc988c253e3ba9be7e7a4fb30748125f474cdc50fdef10b0"} Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.421711 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5vx95" podStartSLOduration=74.421687039 podStartE2EDuration="1m14.421687039s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:10.416974464 +0000 UTC m=+99.579624914" watchObservedRunningTime="2025-11-25 13:48:10.421687039 +0000 UTC m=+99.584337489" Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.422119 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-jqmtw" podStartSLOduration=7.422110589 podStartE2EDuration="7.422110589s" podCreationTimestamp="2025-11-25 13:48:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:10.384951306 +0000 UTC m=+99.547601766" watchObservedRunningTime="2025-11-25 13:48:10.422110589 +0000 UTC m=+99.584761039" Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.433945 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dk89m" event={"ID":"f8caa35c-4401-482b-8c5e-2927fbb50c4d","Type":"ContainerStarted","Data":"657b649d251d5f711563bcc834f512522a0ed1705b7b20c139900e9cd6607956"} Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.434193 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dk89m" Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.444498 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t5vtr" event={"ID":"17279013-fed3-475f-b960-3b9e41116da8","Type":"ContainerStarted","Data":"413abb48cc562b2ad93c34c693e26ec12f25fd58c5f2c768a28ad9a2bee36ee7"} Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.446443 4702 patch_prober.go:28] interesting pod/downloads-7954f5f757-8j9xk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.446515 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8j9xk" podUID="ce097f6c-2894-4417-b66b-0f30d91f3889" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.466637 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.467002 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.472717 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mlxwh" podStartSLOduration=74.472667503 podStartE2EDuration="1m14.472667503s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:10.441851395 +0000 UTC m=+99.604501845" watchObservedRunningTime="2025-11-25 13:48:10.472667503 +0000 UTC m=+99.635317963" Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.473548 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lkx85" podStartSLOduration=74.473541585 podStartE2EDuration="1m14.473541585s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:10.466459651 +0000 UTC m=+99.629110101" watchObservedRunningTime="2025-11-25 13:48:10.473541585 +0000 UTC m=+99.636192035" Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.474831 4702 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-dk89m container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" start-of-body= Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.474912 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dk89m" podUID="f8caa35c-4401-482b-8c5e-2927fbb50c4d" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.480285 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:10 crc kubenswrapper[4702]: E1125 13:48:10.510008 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:11.009980732 +0000 UTC m=+100.172631182 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.511175 4702 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-xsm2s container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.15:8443/livez\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.511217 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" podUID="b619039f-82c2-49ca-9319-75d0883a5a7d" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.15:8443/livez\": dial tcp 10.217.0.15:8443: connect: connection refused" Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.512012 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-rmtvx" Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.513985 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gj88f" Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.547149 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mb6kd" podStartSLOduration=74.547119665 podStartE2EDuration="1m14.547119665s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:10.540491343 +0000 UTC m=+99.703141793" watchObservedRunningTime="2025-11-25 13:48:10.547119665 +0000 UTC m=+99.709770125" Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.582103 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:10 crc kubenswrapper[4702]: E1125 13:48:10.587204 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:11.087177531 +0000 UTC m=+100.249827981 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.589962 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:10 crc kubenswrapper[4702]: E1125 13:48:10.596494 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:11.09647155 +0000 UTC m=+100.259122000 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.628607 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-9bq59" podStartSLOduration=74.62858507 podStartE2EDuration="1m14.62858507s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:10.626042507 +0000 UTC m=+99.788692957" watchObservedRunningTime="2025-11-25 13:48:10.62858507 +0000 UTC m=+99.791235520" Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.700951 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:10 crc kubenswrapper[4702]: E1125 13:48:10.701527 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:11.201503524 +0000 UTC m=+100.364153974 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.724993 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bjl94" podStartSLOduration=74.724974792 podStartE2EDuration="1m14.724974792s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:10.671001733 +0000 UTC m=+99.833652193" watchObservedRunningTime="2025-11-25 13:48:10.724974792 +0000 UTC m=+99.887625242" Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.782081 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-c2xs7" podStartSLOduration=74.782063296 podStartE2EDuration="1m14.782063296s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:10.78059806 +0000 UTC m=+99.943248510" watchObservedRunningTime="2025-11-25 13:48:10.782063296 +0000 UTC m=+99.944713746" Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.782640 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4zsfb" podStartSLOduration=74.78263671 podStartE2EDuration="1m14.78263671s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:10.726576411 +0000 UTC m=+99.889226871" watchObservedRunningTime="2025-11-25 13:48:10.78263671 +0000 UTC m=+99.945287160" Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.802495 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:10 crc kubenswrapper[4702]: E1125 13:48:10.802836 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:11.302825867 +0000 UTC m=+100.465476317 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.831478 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" podStartSLOduration=74.831462172 podStartE2EDuration="1m14.831462172s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:10.830980149 +0000 UTC m=+99.993630599" watchObservedRunningTime="2025-11-25 13:48:10.831462172 +0000 UTC m=+99.994112622" Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.859756 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-8vt69" podStartSLOduration=7.859740607 podStartE2EDuration="7.859740607s" podCreationTimestamp="2025-11-25 13:48:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:10.858441185 +0000 UTC m=+100.021091645" watchObservedRunningTime="2025-11-25 13:48:10.859740607 +0000 UTC m=+100.022391057" Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.899814 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8vr2l" podStartSLOduration=74.899797352 podStartE2EDuration="1m14.899797352s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:10.898760617 +0000 UTC m=+100.061411067" watchObservedRunningTime="2025-11-25 13:48:10.899797352 +0000 UTC m=+100.062447802" Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.903706 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:10 crc kubenswrapper[4702]: E1125 13:48:10.904101 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:11.404088298 +0000 UTC m=+100.566738738 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:10 crc kubenswrapper[4702]: I1125 13:48:10.949927 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dk89m" podStartSLOduration=74.949900385 podStartE2EDuration="1m14.949900385s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:10.946370049 +0000 UTC m=+100.109020509" watchObservedRunningTime="2025-11-25 13:48:10.949900385 +0000 UTC m=+100.112550835" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.006215 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:11 crc kubenswrapper[4702]: E1125 13:48:11.006754 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:11.506732294 +0000 UTC m=+100.669382744 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.107638 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:11 crc kubenswrapper[4702]: E1125 13:48:11.107909 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:11.607861732 +0000 UTC m=+100.770512182 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.108394 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:11 crc kubenswrapper[4702]: E1125 13:48:11.108857 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:11.608849006 +0000 UTC m=+100.771499456 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.210027 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:11 crc kubenswrapper[4702]: E1125 13:48:11.210429 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:11.710370824 +0000 UTC m=+100.873021264 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.210981 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:11 crc kubenswrapper[4702]: E1125 13:48:11.211345 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:11.711326827 +0000 UTC m=+100.873977347 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.312804 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:11 crc kubenswrapper[4702]: E1125 13:48:11.313246 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:11.813208724 +0000 UTC m=+100.975859174 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.359527 4702 patch_prober.go:28] interesting pod/router-default-5444994796-lmq49 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 13:48:11 crc kubenswrapper[4702]: [-]has-synced failed: reason withheld Nov 25 13:48:11 crc kubenswrapper[4702]: [+]process-running ok Nov 25 13:48:11 crc kubenswrapper[4702]: healthz check failed Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.359612 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lmq49" podUID="6d7fc28a-71f2-415c-a99e-76e4ab7fbb70" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.391811 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" podStartSLOduration=76.391789088 podStartE2EDuration="1m16.391789088s" podCreationTimestamp="2025-11-25 13:46:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:11.309918083 +0000 UTC m=+100.472568533" watchObservedRunningTime="2025-11-25 13:48:11.391789088 +0000 UTC m=+100.554439538" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.392363 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t5vtr" podStartSLOduration=75.392356271 podStartE2EDuration="1m15.392356271s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:11.387790479 +0000 UTC m=+100.550440929" watchObservedRunningTime="2025-11-25 13:48:11.392356271 +0000 UTC m=+100.555006721" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.414472 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:11 crc kubenswrapper[4702]: E1125 13:48:11.415048 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:11.915023159 +0000 UTC m=+101.077673819 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.466214 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-v9skf"] Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.467263 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v9skf" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.475269 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.478728 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-5cmt4" event={"ID":"8f55435c-3542-4faf-997d-8a7dad39ce50","Type":"ContainerStarted","Data":"54271f32f9800b7f12d39b164176f5aafbc77e5a2bf928a50d0a9d53dbd40773"} Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.489079 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" event={"ID":"8b821731-b65f-45df-b510-de915e573be1","Type":"ContainerStarted","Data":"075db811eb4ec5ef29459120e9245f5b53e564b496288b2cf95a3e0378e2a013"} Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.495846 4702 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-zqrgn container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.495913 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-zqrgn" podUID="0389edfa-9a86-48df-a9a6-26e2b4d092c4" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.495972 4702 patch_prober.go:28] interesting pod/downloads-7954f5f757-8j9xk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.496034 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8j9xk" podUID="ce097f6c-2894-4417-b66b-0f30d91f3889" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.499151 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t5vtr" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.535321 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:11 crc kubenswrapper[4702]: E1125 13:48:11.535652 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:12.035623726 +0000 UTC m=+101.198274176 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.536352 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:11 crc kubenswrapper[4702]: E1125 13:48:11.536874 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:12.036857016 +0000 UTC m=+101.199507466 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.537696 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-bmb9v" podStartSLOduration=75.537666166 podStartE2EDuration="1m15.537666166s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:11.515430969 +0000 UTC m=+100.678081419" watchObservedRunningTime="2025-11-25 13:48:11.537666166 +0000 UTC m=+100.700316616" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.540953 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dk89m" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.593870 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-j6d5l" podStartSLOduration=75.593844789 podStartE2EDuration="1m15.593844789s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:11.593756706 +0000 UTC m=+100.756407156" watchObservedRunningTime="2025-11-25 13:48:11.593844789 +0000 UTC m=+100.756495229" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.602821 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v9skf"] Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.635889 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pwlcj"] Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.638816 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.639513 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/597c2ecd-d536-4eec-bc02-89c798cbf25c-utilities\") pod \"community-operators-v9skf\" (UID: \"597c2ecd-d536-4eec-bc02-89c798cbf25c\") " pod="openshift-marketplace/community-operators-v9skf" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.639733 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntsgr\" (UniqueName: \"kubernetes.io/projected/597c2ecd-d536-4eec-bc02-89c798cbf25c-kube-api-access-ntsgr\") pod \"community-operators-v9skf\" (UID: \"597c2ecd-d536-4eec-bc02-89c798cbf25c\") " pod="openshift-marketplace/community-operators-v9skf" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.639931 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/597c2ecd-d536-4eec-bc02-89c798cbf25c-catalog-content\") pod \"community-operators-v9skf\" (UID: \"597c2ecd-d536-4eec-bc02-89c798cbf25c\") " pod="openshift-marketplace/community-operators-v9skf" Nov 25 13:48:11 crc kubenswrapper[4702]: E1125 13:48:11.640210 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:12.140182999 +0000 UTC m=+101.302833509 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.640745 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pwlcj" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.659482 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.678222 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pwlcj"] Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.741161 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.741279 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/597c2ecd-d536-4eec-bc02-89c798cbf25c-utilities\") pod \"community-operators-v9skf\" (UID: \"597c2ecd-d536-4eec-bc02-89c798cbf25c\") " pod="openshift-marketplace/community-operators-v9skf" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.741329 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5b4t\" (UniqueName: \"kubernetes.io/projected/66c1bf9b-3a89-427b-9bba-009bdafcfcc8-kube-api-access-l5b4t\") pod \"certified-operators-pwlcj\" (UID: \"66c1bf9b-3a89-427b-9bba-009bdafcfcc8\") " pod="openshift-marketplace/certified-operators-pwlcj" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.741383 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntsgr\" (UniqueName: \"kubernetes.io/projected/597c2ecd-d536-4eec-bc02-89c798cbf25c-kube-api-access-ntsgr\") pod \"community-operators-v9skf\" (UID: \"597c2ecd-d536-4eec-bc02-89c798cbf25c\") " pod="openshift-marketplace/community-operators-v9skf" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.741430 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/597c2ecd-d536-4eec-bc02-89c798cbf25c-catalog-content\") pod \"community-operators-v9skf\" (UID: \"597c2ecd-d536-4eec-bc02-89c798cbf25c\") " pod="openshift-marketplace/community-operators-v9skf" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.741470 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66c1bf9b-3a89-427b-9bba-009bdafcfcc8-utilities\") pod \"certified-operators-pwlcj\" (UID: \"66c1bf9b-3a89-427b-9bba-009bdafcfcc8\") " pod="openshift-marketplace/certified-operators-pwlcj" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.741502 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66c1bf9b-3a89-427b-9bba-009bdafcfcc8-catalog-content\") pod \"certified-operators-pwlcj\" (UID: \"66c1bf9b-3a89-427b-9bba-009bdafcfcc8\") " pod="openshift-marketplace/certified-operators-pwlcj" Nov 25 13:48:11 crc kubenswrapper[4702]: E1125 13:48:11.741894 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:12.241880401 +0000 UTC m=+101.404530851 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.742518 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/597c2ecd-d536-4eec-bc02-89c798cbf25c-utilities\") pod \"community-operators-v9skf\" (UID: \"597c2ecd-d536-4eec-bc02-89c798cbf25c\") " pod="openshift-marketplace/community-operators-v9skf" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.743089 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/597c2ecd-d536-4eec-bc02-89c798cbf25c-catalog-content\") pod \"community-operators-v9skf\" (UID: \"597c2ecd-d536-4eec-bc02-89c798cbf25c\") " pod="openshift-marketplace/community-operators-v9skf" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.810584 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5xb7k"] Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.811824 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5xb7k" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.834365 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntsgr\" (UniqueName: \"kubernetes.io/projected/597c2ecd-d536-4eec-bc02-89c798cbf25c-kube-api-access-ntsgr\") pod \"community-operators-v9skf\" (UID: \"597c2ecd-d536-4eec-bc02-89c798cbf25c\") " pod="openshift-marketplace/community-operators-v9skf" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.842640 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.842960 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5b4t\" (UniqueName: \"kubernetes.io/projected/66c1bf9b-3a89-427b-9bba-009bdafcfcc8-kube-api-access-l5b4t\") pod \"certified-operators-pwlcj\" (UID: \"66c1bf9b-3a89-427b-9bba-009bdafcfcc8\") " pod="openshift-marketplace/certified-operators-pwlcj" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.843027 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66c1bf9b-3a89-427b-9bba-009bdafcfcc8-utilities\") pod \"certified-operators-pwlcj\" (UID: \"66c1bf9b-3a89-427b-9bba-009bdafcfcc8\") " pod="openshift-marketplace/certified-operators-pwlcj" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.843059 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66c1bf9b-3a89-427b-9bba-009bdafcfcc8-catalog-content\") pod \"certified-operators-pwlcj\" (UID: \"66c1bf9b-3a89-427b-9bba-009bdafcfcc8\") " pod="openshift-marketplace/certified-operators-pwlcj" Nov 25 13:48:11 crc kubenswrapper[4702]: E1125 13:48:11.843464 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:12.343440759 +0000 UTC m=+101.506091209 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.843837 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5xb7k"] Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.843932 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66c1bf9b-3a89-427b-9bba-009bdafcfcc8-catalog-content\") pod \"certified-operators-pwlcj\" (UID: \"66c1bf9b-3a89-427b-9bba-009bdafcfcc8\") " pod="openshift-marketplace/certified-operators-pwlcj" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.844078 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66c1bf9b-3a89-427b-9bba-009bdafcfcc8-utilities\") pod \"certified-operators-pwlcj\" (UID: \"66c1bf9b-3a89-427b-9bba-009bdafcfcc8\") " pod="openshift-marketplace/certified-operators-pwlcj" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.882800 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v9skf" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.931099 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5b4t\" (UniqueName: \"kubernetes.io/projected/66c1bf9b-3a89-427b-9bba-009bdafcfcc8-kube-api-access-l5b4t\") pod \"certified-operators-pwlcj\" (UID: \"66c1bf9b-3a89-427b-9bba-009bdafcfcc8\") " pod="openshift-marketplace/certified-operators-pwlcj" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.945613 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6znlx\" (UniqueName: \"kubernetes.io/projected/c48d725b-8394-44f2-90e4-7ba6576fbd77-kube-api-access-6znlx\") pod \"community-operators-5xb7k\" (UID: \"c48d725b-8394-44f2-90e4-7ba6576fbd77\") " pod="openshift-marketplace/community-operators-5xb7k" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.945737 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c48d725b-8394-44f2-90e4-7ba6576fbd77-catalog-content\") pod \"community-operators-5xb7k\" (UID: \"c48d725b-8394-44f2-90e4-7ba6576fbd77\") " pod="openshift-marketplace/community-operators-5xb7k" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.945778 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.945834 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c48d725b-8394-44f2-90e4-7ba6576fbd77-utilities\") pod \"community-operators-5xb7k\" (UID: \"c48d725b-8394-44f2-90e4-7ba6576fbd77\") " pod="openshift-marketplace/community-operators-5xb7k" Nov 25 13:48:11 crc kubenswrapper[4702]: E1125 13:48:11.946215 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:12.446199518 +0000 UTC m=+101.608849968 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.963622 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-7twx2" Nov 25 13:48:11 crc kubenswrapper[4702]: I1125 13:48:11.969331 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pwlcj" Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.004966 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9bfw8"] Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.009500 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9bfw8" Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.036324 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9bfw8"] Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.047087 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.047558 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c48d725b-8394-44f2-90e4-7ba6576fbd77-catalog-content\") pod \"community-operators-5xb7k\" (UID: \"c48d725b-8394-44f2-90e4-7ba6576fbd77\") " pod="openshift-marketplace/community-operators-5xb7k" Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.047752 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c48d725b-8394-44f2-90e4-7ba6576fbd77-utilities\") pod \"community-operators-5xb7k\" (UID: \"c48d725b-8394-44f2-90e4-7ba6576fbd77\") " pod="openshift-marketplace/community-operators-5xb7k" Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.047882 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6znlx\" (UniqueName: \"kubernetes.io/projected/c48d725b-8394-44f2-90e4-7ba6576fbd77-kube-api-access-6znlx\") pod \"community-operators-5xb7k\" (UID: \"c48d725b-8394-44f2-90e4-7ba6576fbd77\") " pod="openshift-marketplace/community-operators-5xb7k" Nov 25 13:48:12 crc kubenswrapper[4702]: E1125 13:48:12.057861 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:12.557820314 +0000 UTC m=+101.720470764 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.058059 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c48d725b-8394-44f2-90e4-7ba6576fbd77-catalog-content\") pod \"community-operators-5xb7k\" (UID: \"c48d725b-8394-44f2-90e4-7ba6576fbd77\") " pod="openshift-marketplace/community-operators-5xb7k" Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.070417 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c48d725b-8394-44f2-90e4-7ba6576fbd77-utilities\") pod \"community-operators-5xb7k\" (UID: \"c48d725b-8394-44f2-90e4-7ba6576fbd77\") " pod="openshift-marketplace/community-operators-5xb7k" Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.110556 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6znlx\" (UniqueName: \"kubernetes.io/projected/c48d725b-8394-44f2-90e4-7ba6576fbd77-kube-api-access-6znlx\") pod \"community-operators-5xb7k\" (UID: \"c48d725b-8394-44f2-90e4-7ba6576fbd77\") " pod="openshift-marketplace/community-operators-5xb7k" Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.151713 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5xb7k" Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.153375 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3babd873-f61e-40d1-a42f-8a2f32f3ca40-utilities\") pod \"certified-operators-9bfw8\" (UID: \"3babd873-f61e-40d1-a42f-8a2f32f3ca40\") " pod="openshift-marketplace/certified-operators-9bfw8" Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.153439 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3babd873-f61e-40d1-a42f-8a2f32f3ca40-catalog-content\") pod \"certified-operators-9bfw8\" (UID: \"3babd873-f61e-40d1-a42f-8a2f32f3ca40\") " pod="openshift-marketplace/certified-operators-9bfw8" Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.153457 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr26k\" (UniqueName: \"kubernetes.io/projected/3babd873-f61e-40d1-a42f-8a2f32f3ca40-kube-api-access-qr26k\") pod \"certified-operators-9bfw8\" (UID: \"3babd873-f61e-40d1-a42f-8a2f32f3ca40\") " pod="openshift-marketplace/certified-operators-9bfw8" Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.153483 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:12 crc kubenswrapper[4702]: E1125 13:48:12.153739 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:12.653727564 +0000 UTC m=+101.816378014 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.263146 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.263646 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3babd873-f61e-40d1-a42f-8a2f32f3ca40-catalog-content\") pod \"certified-operators-9bfw8\" (UID: \"3babd873-f61e-40d1-a42f-8a2f32f3ca40\") " pod="openshift-marketplace/certified-operators-9bfw8" Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.263700 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr26k\" (UniqueName: \"kubernetes.io/projected/3babd873-f61e-40d1-a42f-8a2f32f3ca40-kube-api-access-qr26k\") pod \"certified-operators-9bfw8\" (UID: \"3babd873-f61e-40d1-a42f-8a2f32f3ca40\") " pod="openshift-marketplace/certified-operators-9bfw8" Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.263867 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3babd873-f61e-40d1-a42f-8a2f32f3ca40-utilities\") pod \"certified-operators-9bfw8\" (UID: \"3babd873-f61e-40d1-a42f-8a2f32f3ca40\") " pod="openshift-marketplace/certified-operators-9bfw8" Nov 25 13:48:12 crc kubenswrapper[4702]: E1125 13:48:12.264889 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:12.764865458 +0000 UTC m=+101.927515898 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.265030 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3babd873-f61e-40d1-a42f-8a2f32f3ca40-catalog-content\") pod \"certified-operators-9bfw8\" (UID: \"3babd873-f61e-40d1-a42f-8a2f32f3ca40\") " pod="openshift-marketplace/certified-operators-9bfw8" Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.265078 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3babd873-f61e-40d1-a42f-8a2f32f3ca40-utilities\") pod \"certified-operators-9bfw8\" (UID: \"3babd873-f61e-40d1-a42f-8a2f32f3ca40\") " pod="openshift-marketplace/certified-operators-9bfw8" Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.322940 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr26k\" (UniqueName: \"kubernetes.io/projected/3babd873-f61e-40d1-a42f-8a2f32f3ca40-kube-api-access-qr26k\") pod \"certified-operators-9bfw8\" (UID: \"3babd873-f61e-40d1-a42f-8a2f32f3ca40\") " pod="openshift-marketplace/certified-operators-9bfw8" Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.328699 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9bfw8" Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.362952 4702 patch_prober.go:28] interesting pod/router-default-5444994796-lmq49 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 13:48:12 crc kubenswrapper[4702]: [-]has-synced failed: reason withheld Nov 25 13:48:12 crc kubenswrapper[4702]: [+]process-running ok Nov 25 13:48:12 crc kubenswrapper[4702]: healthz check failed Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.363388 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lmq49" podUID="6d7fc28a-71f2-415c-a99e-76e4ab7fbb70" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.365142 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:12 crc kubenswrapper[4702]: E1125 13:48:12.365687 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:12.865670138 +0000 UTC m=+102.028320588 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.467957 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:12 crc kubenswrapper[4702]: E1125 13:48:12.468581 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:12.96856374 +0000 UTC m=+102.131214190 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.479468 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t5vtr" Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.526824 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-5cmt4" event={"ID":"8f55435c-3542-4faf-997d-8a7dad39ce50","Type":"ContainerStarted","Data":"48e9050fe1a43198bfd2af520c095b82f5a7b27e734d624c2a3478f665bfa85c"} Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.553595 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-zqrgn" Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.564767 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v9skf"] Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.574881 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:12 crc kubenswrapper[4702]: E1125 13:48:12.575438 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:13.075421309 +0000 UTC m=+102.238071759 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.675893 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:12 crc kubenswrapper[4702]: E1125 13:48:12.677570 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:13.177542882 +0000 UTC m=+102.340193372 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.780371 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:12 crc kubenswrapper[4702]: E1125 13:48:12.780721 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:13.28070704 +0000 UTC m=+102.443357490 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.844841 4702 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.884686 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:12 crc kubenswrapper[4702]: E1125 13:48:12.884932 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:13.384902603 +0000 UTC m=+102.547553053 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.885200 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:12 crc kubenswrapper[4702]: E1125 13:48:12.886470 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:13.386451801 +0000 UTC m=+102.549102441 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.987461 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:12 crc kubenswrapper[4702]: E1125 13:48:12.987705 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:13.487669492 +0000 UTC m=+102.650319952 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:12 crc kubenswrapper[4702]: I1125 13:48:12.987849 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:12 crc kubenswrapper[4702]: E1125 13:48:12.988298 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:13.488288587 +0000 UTC m=+102.650939037 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.045335 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pwlcj"] Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.098756 4702 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-25T13:48:12.845101124Z","Handler":null,"Name":""} Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.099192 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:13 crc kubenswrapper[4702]: E1125 13:48:13.099280 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 13:48:13.599256037 +0000 UTC m=+102.761906487 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.099707 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:13 crc kubenswrapper[4702]: E1125 13:48:13.100049 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 13:48:13.600036787 +0000 UTC m=+102.762687237 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vlz75" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.150859 4702 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.150922 4702 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.174914 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9bfw8"] Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.203817 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.240000 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.243160 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5xb7k"] Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.306982 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.357468 4702 patch_prober.go:28] interesting pod/router-default-5444994796-lmq49 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 13:48:13 crc kubenswrapper[4702]: [-]has-synced failed: reason withheld Nov 25 13:48:13 crc kubenswrapper[4702]: [+]process-running ok Nov 25 13:48:13 crc kubenswrapper[4702]: healthz check failed Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.357530 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lmq49" podUID="6d7fc28a-71f2-415c-a99e-76e4ab7fbb70" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.357945 4702 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.358003 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.443675 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vlz75\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.452293 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.550876 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5xb7k" event={"ID":"c48d725b-8394-44f2-90e4-7ba6576fbd77","Type":"ContainerStarted","Data":"2d9de4ac3da08660b0236131a82991bc738884e9d4d8764f178aedc43125e5cf"} Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.553102 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9bfw8" event={"ID":"3babd873-f61e-40d1-a42f-8a2f32f3ca40","Type":"ContainerStarted","Data":"cd3f34747ba3c2c05a0acdea6cd6f1f5f8db42f2bdd2a5055b3110d74ba9deb1"} Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.579288 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-5cmt4" event={"ID":"8f55435c-3542-4faf-997d-8a7dad39ce50","Type":"ContainerStarted","Data":"3a2d84e9341347464cf38f548bd8864cb1ab6a4ebb620df099988d0428e419de"} Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.590422 4702 generic.go:334] "Generic (PLEG): container finished" podID="597c2ecd-d536-4eec-bc02-89c798cbf25c" containerID="e7efc31fb65168ba89ae8b20e86503b7019eae41added19d7492b0e7d9058e5c" exitCode=0 Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.590494 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v9skf" event={"ID":"597c2ecd-d536-4eec-bc02-89c798cbf25c","Type":"ContainerDied","Data":"e7efc31fb65168ba89ae8b20e86503b7019eae41added19d7492b0e7d9058e5c"} Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.590522 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v9skf" event={"ID":"597c2ecd-d536-4eec-bc02-89c798cbf25c","Type":"ContainerStarted","Data":"77323b03247449a75a17a0a44b6b13d98acc7d342ad48da6c4b13244d733446c"} Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.593388 4702 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.593486 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pwlcj" event={"ID":"66c1bf9b-3a89-427b-9bba-009bdafcfcc8","Type":"ContainerStarted","Data":"582d5036f631fbdb494a4941759f3a9a42629120c463523af17627105cbe0d8e"} Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.614627 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jq6gh"] Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.615863 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jq6gh" Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.620649 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.645604 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-5cmt4" podStartSLOduration=10.645572848 podStartE2EDuration="10.645572848s" podCreationTimestamp="2025-11-25 13:48:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:13.632907797 +0000 UTC m=+102.795558247" watchObservedRunningTime="2025-11-25 13:48:13.645572848 +0000 UTC m=+102.808223298" Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.647919 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jq6gh"] Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.715161 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5b74e23-72c6-4278-9dc7-909f528d344b-catalog-content\") pod \"redhat-marketplace-jq6gh\" (UID: \"f5b74e23-72c6-4278-9dc7-909f528d344b\") " pod="openshift-marketplace/redhat-marketplace-jq6gh" Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.715379 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swzxn\" (UniqueName: \"kubernetes.io/projected/f5b74e23-72c6-4278-9dc7-909f528d344b-kube-api-access-swzxn\") pod \"redhat-marketplace-jq6gh\" (UID: \"f5b74e23-72c6-4278-9dc7-909f528d344b\") " pod="openshift-marketplace/redhat-marketplace-jq6gh" Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.715468 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5b74e23-72c6-4278-9dc7-909f528d344b-utilities\") pod \"redhat-marketplace-jq6gh\" (UID: \"f5b74e23-72c6-4278-9dc7-909f528d344b\") " pod="openshift-marketplace/redhat-marketplace-jq6gh" Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.818961 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.824620 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5b74e23-72c6-4278-9dc7-909f528d344b-catalog-content\") pod \"redhat-marketplace-jq6gh\" (UID: \"f5b74e23-72c6-4278-9dc7-909f528d344b\") " pod="openshift-marketplace/redhat-marketplace-jq6gh" Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.824739 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swzxn\" (UniqueName: \"kubernetes.io/projected/f5b74e23-72c6-4278-9dc7-909f528d344b-kube-api-access-swzxn\") pod \"redhat-marketplace-jq6gh\" (UID: \"f5b74e23-72c6-4278-9dc7-909f528d344b\") " pod="openshift-marketplace/redhat-marketplace-jq6gh" Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.824794 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5b74e23-72c6-4278-9dc7-909f528d344b-utilities\") pod \"redhat-marketplace-jq6gh\" (UID: \"f5b74e23-72c6-4278-9dc7-909f528d344b\") " pod="openshift-marketplace/redhat-marketplace-jq6gh" Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.825332 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5b74e23-72c6-4278-9dc7-909f528d344b-utilities\") pod \"redhat-marketplace-jq6gh\" (UID: \"f5b74e23-72c6-4278-9dc7-909f528d344b\") " pod="openshift-marketplace/redhat-marketplace-jq6gh" Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.825632 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5b74e23-72c6-4278-9dc7-909f528d344b-catalog-content\") pod \"redhat-marketplace-jq6gh\" (UID: \"f5b74e23-72c6-4278-9dc7-909f528d344b\") " pod="openshift-marketplace/redhat-marketplace-jq6gh" Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.872828 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swzxn\" (UniqueName: \"kubernetes.io/projected/f5b74e23-72c6-4278-9dc7-909f528d344b-kube-api-access-swzxn\") pod \"redhat-marketplace-jq6gh\" (UID: \"f5b74e23-72c6-4278-9dc7-909f528d344b\") " pod="openshift-marketplace/redhat-marketplace-jq6gh" Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.902568 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vlz75"] Nov 25 13:48:13 crc kubenswrapper[4702]: W1125 13:48:13.914340 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda70d6c59_6b85_4950_8e30_1938da7a01e8.slice/crio-a3443723263d837e8e18ee56c1ff63c829f60df2aeab2f735bf225cb08c91c98 WatchSource:0}: Error finding container a3443723263d837e8e18ee56c1ff63c829f60df2aeab2f735bf225cb08c91c98: Status 404 returned error can't find the container with id a3443723263d837e8e18ee56c1ff63c829f60df2aeab2f735bf225cb08c91c98 Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.952628 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jq6gh" Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.990169 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xm8tp"] Nov 25 13:48:13 crc kubenswrapper[4702]: I1125 13:48:13.991105 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xm8tp" Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.073366 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xm8tp"] Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.129344 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hhjf\" (UniqueName: \"kubernetes.io/projected/a78d7c57-6155-429a-bbcf-168c3fa26e2e-kube-api-access-6hhjf\") pod \"redhat-marketplace-xm8tp\" (UID: \"a78d7c57-6155-429a-bbcf-168c3fa26e2e\") " pod="openshift-marketplace/redhat-marketplace-xm8tp" Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.129763 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a78d7c57-6155-429a-bbcf-168c3fa26e2e-utilities\") pod \"redhat-marketplace-xm8tp\" (UID: \"a78d7c57-6155-429a-bbcf-168c3fa26e2e\") " pod="openshift-marketplace/redhat-marketplace-xm8tp" Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.129864 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a78d7c57-6155-429a-bbcf-168c3fa26e2e-catalog-content\") pod \"redhat-marketplace-xm8tp\" (UID: \"a78d7c57-6155-429a-bbcf-168c3fa26e2e\") " pod="openshift-marketplace/redhat-marketplace-xm8tp" Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.230900 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hhjf\" (UniqueName: \"kubernetes.io/projected/a78d7c57-6155-429a-bbcf-168c3fa26e2e-kube-api-access-6hhjf\") pod \"redhat-marketplace-xm8tp\" (UID: \"a78d7c57-6155-429a-bbcf-168c3fa26e2e\") " pod="openshift-marketplace/redhat-marketplace-xm8tp" Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.230991 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a78d7c57-6155-429a-bbcf-168c3fa26e2e-utilities\") pod \"redhat-marketplace-xm8tp\" (UID: \"a78d7c57-6155-429a-bbcf-168c3fa26e2e\") " pod="openshift-marketplace/redhat-marketplace-xm8tp" Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.231030 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a78d7c57-6155-429a-bbcf-168c3fa26e2e-catalog-content\") pod \"redhat-marketplace-xm8tp\" (UID: \"a78d7c57-6155-429a-bbcf-168c3fa26e2e\") " pod="openshift-marketplace/redhat-marketplace-xm8tp" Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.231574 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a78d7c57-6155-429a-bbcf-168c3fa26e2e-catalog-content\") pod \"redhat-marketplace-xm8tp\" (UID: \"a78d7c57-6155-429a-bbcf-168c3fa26e2e\") " pod="openshift-marketplace/redhat-marketplace-xm8tp" Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.236887 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a78d7c57-6155-429a-bbcf-168c3fa26e2e-utilities\") pod \"redhat-marketplace-xm8tp\" (UID: \"a78d7c57-6155-429a-bbcf-168c3fa26e2e\") " pod="openshift-marketplace/redhat-marketplace-xm8tp" Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.292303 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hhjf\" (UniqueName: \"kubernetes.io/projected/a78d7c57-6155-429a-bbcf-168c3fa26e2e-kube-api-access-6hhjf\") pod \"redhat-marketplace-xm8tp\" (UID: \"a78d7c57-6155-429a-bbcf-168c3fa26e2e\") " pod="openshift-marketplace/redhat-marketplace-xm8tp" Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.328224 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xm8tp" Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.356530 4702 patch_prober.go:28] interesting pod/router-default-5444994796-lmq49 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 13:48:14 crc kubenswrapper[4702]: [-]has-synced failed: reason withheld Nov 25 13:48:14 crc kubenswrapper[4702]: [+]process-running ok Nov 25 13:48:14 crc kubenswrapper[4702]: healthz check failed Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.356604 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lmq49" podUID="6d7fc28a-71f2-415c-a99e-76e4ab7fbb70" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.392611 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jq6gh"] Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.537608 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/88665994-5305-4de7-8b69-e38b38d5f87c-metrics-certs\") pod \"network-metrics-daemon-fm2cr\" (UID: \"88665994-5305-4de7-8b69-e38b38d5f87c\") " pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.542902 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/88665994-5305-4de7-8b69-e38b38d5f87c-metrics-certs\") pod \"network-metrics-daemon-fm2cr\" (UID: \"88665994-5305-4de7-8b69-e38b38d5f87c\") " pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.590638 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ppzc7"] Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.591611 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ppzc7" Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.593766 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.601786 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jq6gh" event={"ID":"f5b74e23-72c6-4278-9dc7-909f528d344b","Type":"ContainerStarted","Data":"d6426bd7473101ce6ed7cf2ab5c1d3d453559de0441ee92b877072211873e9d4"} Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.603312 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ppzc7"] Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.604968 4702 generic.go:334] "Generic (PLEG): container finished" podID="c48d725b-8394-44f2-90e4-7ba6576fbd77" containerID="ef6eb6b37fb585832cecf57b30ced7e5104bb925b10b7e2f9cdf1ea63fee1c39" exitCode=0 Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.605105 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5xb7k" event={"ID":"c48d725b-8394-44f2-90e4-7ba6576fbd77","Type":"ContainerDied","Data":"ef6eb6b37fb585832cecf57b30ced7e5104bb925b10b7e2f9cdf1ea63fee1c39"} Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.607223 4702 generic.go:334] "Generic (PLEG): container finished" podID="3babd873-f61e-40d1-a42f-8a2f32f3ca40" containerID="996fb999cb3152d20808978aa1e2361941f291157495c8f7203d464c24ea18ca" exitCode=0 Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.607462 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9bfw8" event={"ID":"3babd873-f61e-40d1-a42f-8a2f32f3ca40","Type":"ContainerDied","Data":"996fb999cb3152d20808978aa1e2361941f291157495c8f7203d464c24ea18ca"} Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.611610 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" event={"ID":"a70d6c59-6b85-4950-8e30-1938da7a01e8","Type":"ContainerStarted","Data":"2c3c42ca85810566c17151205d72d3310991d375f0e4e80d8686edd5c6695b87"} Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.611642 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" event={"ID":"a70d6c59-6b85-4950-8e30-1938da7a01e8","Type":"ContainerStarted","Data":"a3443723263d837e8e18ee56c1ff63c829f60df2aeab2f735bf225cb08c91c98"} Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.611851 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.613868 4702 generic.go:334] "Generic (PLEG): container finished" podID="66c1bf9b-3a89-427b-9bba-009bdafcfcc8" containerID="d18368f3544c7485a747ab6635a3dabf62583e82e595fd3004b7da8648b97179" exitCode=0 Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.615335 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pwlcj" event={"ID":"66c1bf9b-3a89-427b-9bba-009bdafcfcc8","Type":"ContainerDied","Data":"d18368f3544c7485a747ab6635a3dabf62583e82e595fd3004b7da8648b97179"} Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.638140 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-c2xs7" Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.643507 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xm8tp"] Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.689871 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" podStartSLOduration=78.689843681 podStartE2EDuration="1m18.689843681s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:14.672614858 +0000 UTC m=+103.835265308" watchObservedRunningTime="2025-11-25 13:48:14.689843681 +0000 UTC m=+103.852494131" Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.741920 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68f8eda2-8222-4ecb-a2d0-316ed8398d58-catalog-content\") pod \"redhat-operators-ppzc7\" (UID: \"68f8eda2-8222-4ecb-a2d0-316ed8398d58\") " pod="openshift-marketplace/redhat-operators-ppzc7" Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.741978 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68f8eda2-8222-4ecb-a2d0-316ed8398d58-utilities\") pod \"redhat-operators-ppzc7\" (UID: \"68f8eda2-8222-4ecb-a2d0-316ed8398d58\") " pod="openshift-marketplace/redhat-operators-ppzc7" Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.742477 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzwpq\" (UniqueName: \"kubernetes.io/projected/68f8eda2-8222-4ecb-a2d0-316ed8398d58-kube-api-access-jzwpq\") pod \"redhat-operators-ppzc7\" (UID: \"68f8eda2-8222-4ecb-a2d0-316ed8398d58\") " pod="openshift-marketplace/redhat-operators-ppzc7" Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.743304 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fm2cr" Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.844252 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzwpq\" (UniqueName: \"kubernetes.io/projected/68f8eda2-8222-4ecb-a2d0-316ed8398d58-kube-api-access-jzwpq\") pod \"redhat-operators-ppzc7\" (UID: \"68f8eda2-8222-4ecb-a2d0-316ed8398d58\") " pod="openshift-marketplace/redhat-operators-ppzc7" Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.844350 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68f8eda2-8222-4ecb-a2d0-316ed8398d58-catalog-content\") pod \"redhat-operators-ppzc7\" (UID: \"68f8eda2-8222-4ecb-a2d0-316ed8398d58\") " pod="openshift-marketplace/redhat-operators-ppzc7" Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.844381 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68f8eda2-8222-4ecb-a2d0-316ed8398d58-utilities\") pod \"redhat-operators-ppzc7\" (UID: \"68f8eda2-8222-4ecb-a2d0-316ed8398d58\") " pod="openshift-marketplace/redhat-operators-ppzc7" Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.845101 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68f8eda2-8222-4ecb-a2d0-316ed8398d58-utilities\") pod \"redhat-operators-ppzc7\" (UID: \"68f8eda2-8222-4ecb-a2d0-316ed8398d58\") " pod="openshift-marketplace/redhat-operators-ppzc7" Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.845126 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68f8eda2-8222-4ecb-a2d0-316ed8398d58-catalog-content\") pod \"redhat-operators-ppzc7\" (UID: \"68f8eda2-8222-4ecb-a2d0-316ed8398d58\") " pod="openshift-marketplace/redhat-operators-ppzc7" Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.870762 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzwpq\" (UniqueName: \"kubernetes.io/projected/68f8eda2-8222-4ecb-a2d0-316ed8398d58-kube-api-access-jzwpq\") pod \"redhat-operators-ppzc7\" (UID: \"68f8eda2-8222-4ecb-a2d0-316ed8398d58\") " pod="openshift-marketplace/redhat-operators-ppzc7" Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.910913 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ppzc7" Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.978300 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-fm2cr"] Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.993972 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rh58q"] Nov 25 13:48:14 crc kubenswrapper[4702]: I1125 13:48:14.995289 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rh58q" Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.008327 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rh58q"] Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.124968 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ppzc7"] Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.140087 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.141369 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.158384 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.158582 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.159595 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wncr\" (UniqueName: \"kubernetes.io/projected/75d44e5f-cb22-4c92-8695-b9b8128d8724-kube-api-access-4wncr\") pod \"redhat-operators-rh58q\" (UID: \"75d44e5f-cb22-4c92-8695-b9b8128d8724\") " pod="openshift-marketplace/redhat-operators-rh58q" Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.159666 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75d44e5f-cb22-4c92-8695-b9b8128d8724-utilities\") pod \"redhat-operators-rh58q\" (UID: \"75d44e5f-cb22-4c92-8695-b9b8128d8724\") " pod="openshift-marketplace/redhat-operators-rh58q" Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.159823 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75d44e5f-cb22-4c92-8695-b9b8128d8724-catalog-content\") pod \"redhat-operators-rh58q\" (UID: \"75d44e5f-cb22-4c92-8695-b9b8128d8724\") " pod="openshift-marketplace/redhat-operators-rh58q" Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.160174 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.260917 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75d44e5f-cb22-4c92-8695-b9b8128d8724-catalog-content\") pod \"redhat-operators-rh58q\" (UID: \"75d44e5f-cb22-4c92-8695-b9b8128d8724\") " pod="openshift-marketplace/redhat-operators-rh58q" Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.260961 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9ab03842-ddff-42ae-8378-8c5a6cc4fb35-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9ab03842-ddff-42ae-8378-8c5a6cc4fb35\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.261081 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wncr\" (UniqueName: \"kubernetes.io/projected/75d44e5f-cb22-4c92-8695-b9b8128d8724-kube-api-access-4wncr\") pod \"redhat-operators-rh58q\" (UID: \"75d44e5f-cb22-4c92-8695-b9b8128d8724\") " pod="openshift-marketplace/redhat-operators-rh58q" Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.261104 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75d44e5f-cb22-4c92-8695-b9b8128d8724-utilities\") pod \"redhat-operators-rh58q\" (UID: \"75d44e5f-cb22-4c92-8695-b9b8128d8724\") " pod="openshift-marketplace/redhat-operators-rh58q" Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.261128 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9ab03842-ddff-42ae-8378-8c5a6cc4fb35-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9ab03842-ddff-42ae-8378-8c5a6cc4fb35\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.261783 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75d44e5f-cb22-4c92-8695-b9b8128d8724-catalog-content\") pod \"redhat-operators-rh58q\" (UID: \"75d44e5f-cb22-4c92-8695-b9b8128d8724\") " pod="openshift-marketplace/redhat-operators-rh58q" Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.262283 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75d44e5f-cb22-4c92-8695-b9b8128d8724-utilities\") pod \"redhat-operators-rh58q\" (UID: \"75d44e5f-cb22-4c92-8695-b9b8128d8724\") " pod="openshift-marketplace/redhat-operators-rh58q" Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.284779 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wncr\" (UniqueName: \"kubernetes.io/projected/75d44e5f-cb22-4c92-8695-b9b8128d8724-kube-api-access-4wncr\") pod \"redhat-operators-rh58q\" (UID: \"75d44e5f-cb22-4c92-8695-b9b8128d8724\") " pod="openshift-marketplace/redhat-operators-rh58q" Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.321923 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rh58q" Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.356198 4702 patch_prober.go:28] interesting pod/router-default-5444994796-lmq49 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 13:48:15 crc kubenswrapper[4702]: [-]has-synced failed: reason withheld Nov 25 13:48:15 crc kubenswrapper[4702]: [+]process-running ok Nov 25 13:48:15 crc kubenswrapper[4702]: healthz check failed Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.356274 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lmq49" podUID="6d7fc28a-71f2-415c-a99e-76e4ab7fbb70" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.362899 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9ab03842-ddff-42ae-8378-8c5a6cc4fb35-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9ab03842-ddff-42ae-8378-8c5a6cc4fb35\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.362996 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9ab03842-ddff-42ae-8378-8c5a6cc4fb35-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9ab03842-ddff-42ae-8378-8c5a6cc4fb35\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.363353 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9ab03842-ddff-42ae-8378-8c5a6cc4fb35-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9ab03842-ddff-42ae-8378-8c5a6cc4fb35\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.369156 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.370219 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.385133 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9ab03842-ddff-42ae-8378-8c5a6cc4fb35-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9ab03842-ddff-42ae-8378-8c5a6cc4fb35\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.389489 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.410630 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-mbmx2" Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.412339 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-mbmx2" Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.415510 4702 patch_prober.go:28] interesting pod/console-f9d7485db-mbmx2 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.415547 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-mbmx2" podUID="dd97957f-e7df-48c2-b23f-b1101b786766" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.466649 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.473303 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xsm2s" Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.574976 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.596504 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rh58q"] Nov 25 13:48:15 crc kubenswrapper[4702]: W1125 13:48:15.622415 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75d44e5f_cb22_4c92_8695_b9b8128d8724.slice/crio-4aebbab9169ca54a57e4034f63a82d7d371f38cf0ac5666265cc838c171620ce WatchSource:0}: Error finding container 4aebbab9169ca54a57e4034f63a82d7d371f38cf0ac5666265cc838c171620ce: Status 404 returned error can't find the container with id 4aebbab9169ca54a57e4034f63a82d7d371f38cf0ac5666265cc838c171620ce Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.624827 4702 generic.go:334] "Generic (PLEG): container finished" podID="8cb1a228-426f-41cc-a15f-01dfb84a35cd" containerID="c8fc47a36657d5c1b5bd76bdeca4e43395474e5ef3305ab606d2c7edfe968fa0" exitCode=0 Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.624963 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401305-j5rqf" event={"ID":"8cb1a228-426f-41cc-a15f-01dfb84a35cd","Type":"ContainerDied","Data":"c8fc47a36657d5c1b5bd76bdeca4e43395474e5ef3305ab606d2c7edfe968fa0"} Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.631456 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-fm2cr" event={"ID":"88665994-5305-4de7-8b69-e38b38d5f87c","Type":"ContainerStarted","Data":"29163cf6e4187b85360d7178e1236d499ec7a54695f635181ca3c1cc12e83d06"} Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.631525 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-fm2cr" event={"ID":"88665994-5305-4de7-8b69-e38b38d5f87c","Type":"ContainerStarted","Data":"e5dd231856982a296d2fc534c3cf9db965a160858f961e1050a458c287a2adfb"} Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.636506 4702 generic.go:334] "Generic (PLEG): container finished" podID="f5b74e23-72c6-4278-9dc7-909f528d344b" containerID="052bbacfe3923529d43f8faf0dcca91bd7695404ebd9b87324946c534e265603" exitCode=0 Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.637453 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jq6gh" event={"ID":"f5b74e23-72c6-4278-9dc7-909f528d344b","Type":"ContainerDied","Data":"052bbacfe3923529d43f8faf0dcca91bd7695404ebd9b87324946c534e265603"} Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.638714 4702 patch_prober.go:28] interesting pod/downloads-7954f5f757-8j9xk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.638728 4702 patch_prober.go:28] interesting pod/downloads-7954f5f757-8j9xk container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.638770 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8j9xk" podUID="ce097f6c-2894-4417-b66b-0f30d91f3889" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.638822 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-8j9xk" podUID="ce097f6c-2894-4417-b66b-0f30d91f3889" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.671082 4702 generic.go:334] "Generic (PLEG): container finished" podID="a78d7c57-6155-429a-bbcf-168c3fa26e2e" containerID="103f8ed0fc0c0650810aee82d33999bee26212a92d3a64ba13b61a471ceed5ed" exitCode=0 Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.671175 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xm8tp" event={"ID":"a78d7c57-6155-429a-bbcf-168c3fa26e2e","Type":"ContainerDied","Data":"103f8ed0fc0c0650810aee82d33999bee26212a92d3a64ba13b61a471ceed5ed"} Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.671208 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xm8tp" event={"ID":"a78d7c57-6155-429a-bbcf-168c3fa26e2e","Type":"ContainerStarted","Data":"8ea4ebe9c214764646b49ce4213144f6b300e8a3ee5ba3819c2c0594266ba648"} Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.674618 4702 generic.go:334] "Generic (PLEG): container finished" podID="68f8eda2-8222-4ecb-a2d0-316ed8398d58" containerID="ee2a8af9ab5feb0f3a51eb3cb536a2a8c229072a07438479512f4b80f5f7dcdb" exitCode=0 Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.677345 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ppzc7" event={"ID":"68f8eda2-8222-4ecb-a2d0-316ed8398d58","Type":"ContainerDied","Data":"ee2a8af9ab5feb0f3a51eb3cb536a2a8c229072a07438479512f4b80f5f7dcdb"} Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.678226 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ppzc7" event={"ID":"68f8eda2-8222-4ecb-a2d0-316ed8398d58","Type":"ContainerStarted","Data":"ed80ea5f96d6c387b4069090ffeb929bd2ca6dd1b70449cc88264b8a31c9fdb4"} Nov 25 13:48:15 crc kubenswrapper[4702]: I1125 13:48:15.690176 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-pvqd2" Nov 25 13:48:16 crc kubenswrapper[4702]: I1125 13:48:16.012916 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 13:48:16 crc kubenswrapper[4702]: I1125 13:48:16.064252 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:48:16 crc kubenswrapper[4702]: I1125 13:48:16.351445 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-lmq49" Nov 25 13:48:16 crc kubenswrapper[4702]: I1125 13:48:16.354653 4702 patch_prober.go:28] interesting pod/router-default-5444994796-lmq49 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 13:48:16 crc kubenswrapper[4702]: [-]has-synced failed: reason withheld Nov 25 13:48:16 crc kubenswrapper[4702]: [+]process-running ok Nov 25 13:48:16 crc kubenswrapper[4702]: healthz check failed Nov 25 13:48:16 crc kubenswrapper[4702]: I1125 13:48:16.354697 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lmq49" podUID="6d7fc28a-71f2-415c-a99e-76e4ab7fbb70" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 13:48:16 crc kubenswrapper[4702]: I1125 13:48:16.687082 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-fm2cr" event={"ID":"88665994-5305-4de7-8b69-e38b38d5f87c","Type":"ContainerStarted","Data":"33bf6073cab7a32e278cd35c01e5c91e20df98105255a3259294013e5f78d1a7"} Nov 25 13:48:16 crc kubenswrapper[4702]: I1125 13:48:16.690832 4702 generic.go:334] "Generic (PLEG): container finished" podID="75d44e5f-cb22-4c92-8695-b9b8128d8724" containerID="2349a88076ca64f27e2459432bcc42ee0349d91fd7626e82f04950b44b9cbc8e" exitCode=0 Nov 25 13:48:16 crc kubenswrapper[4702]: I1125 13:48:16.690921 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rh58q" event={"ID":"75d44e5f-cb22-4c92-8695-b9b8128d8724","Type":"ContainerDied","Data":"2349a88076ca64f27e2459432bcc42ee0349d91fd7626e82f04950b44b9cbc8e"} Nov 25 13:48:16 crc kubenswrapper[4702]: I1125 13:48:16.690944 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rh58q" event={"ID":"75d44e5f-cb22-4c92-8695-b9b8128d8724","Type":"ContainerStarted","Data":"4aebbab9169ca54a57e4034f63a82d7d371f38cf0ac5666265cc838c171620ce"} Nov 25 13:48:16 crc kubenswrapper[4702]: I1125 13:48:16.693892 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"9ab03842-ddff-42ae-8378-8c5a6cc4fb35","Type":"ContainerStarted","Data":"3071ed6707bc47c77c110035c8e4a4546921639f6e2fd1b9b10fb0b28c0c8308"} Nov 25 13:48:16 crc kubenswrapper[4702]: I1125 13:48:16.693954 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"9ab03842-ddff-42ae-8378-8c5a6cc4fb35","Type":"ContainerStarted","Data":"17565ea5286d9300ba7085d138617f4441058f5e30e93087c018d7d80f6c2118"} Nov 25 13:48:16 crc kubenswrapper[4702]: I1125 13:48:16.706479 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-fm2cr" podStartSLOduration=80.706461228 podStartE2EDuration="1m20.706461228s" podCreationTimestamp="2025-11-25 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:16.705411162 +0000 UTC m=+105.868061612" watchObservedRunningTime="2025-11-25 13:48:16.706461228 +0000 UTC m=+105.869111678" Nov 25 13:48:16 crc kubenswrapper[4702]: I1125 13:48:16.722537 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=1.722521553 podStartE2EDuration="1.722521553s" podCreationTimestamp="2025-11-25 13:48:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:16.720032052 +0000 UTC m=+105.882682502" watchObservedRunningTime="2025-11-25 13:48:16.722521553 +0000 UTC m=+105.885172003" Nov 25 13:48:17 crc kubenswrapper[4702]: I1125 13:48:17.006766 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401305-j5rqf" Nov 25 13:48:17 crc kubenswrapper[4702]: I1125 13:48:17.122679 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgbkn\" (UniqueName: \"kubernetes.io/projected/8cb1a228-426f-41cc-a15f-01dfb84a35cd-kube-api-access-jgbkn\") pod \"8cb1a228-426f-41cc-a15f-01dfb84a35cd\" (UID: \"8cb1a228-426f-41cc-a15f-01dfb84a35cd\") " Nov 25 13:48:17 crc kubenswrapper[4702]: I1125 13:48:17.122813 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8cb1a228-426f-41cc-a15f-01dfb84a35cd-config-volume\") pod \"8cb1a228-426f-41cc-a15f-01dfb84a35cd\" (UID: \"8cb1a228-426f-41cc-a15f-01dfb84a35cd\") " Nov 25 13:48:17 crc kubenswrapper[4702]: I1125 13:48:17.122902 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8cb1a228-426f-41cc-a15f-01dfb84a35cd-secret-volume\") pod \"8cb1a228-426f-41cc-a15f-01dfb84a35cd\" (UID: \"8cb1a228-426f-41cc-a15f-01dfb84a35cd\") " Nov 25 13:48:17 crc kubenswrapper[4702]: I1125 13:48:17.127834 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cb1a228-426f-41cc-a15f-01dfb84a35cd-config-volume" (OuterVolumeSpecName: "config-volume") pod "8cb1a228-426f-41cc-a15f-01dfb84a35cd" (UID: "8cb1a228-426f-41cc-a15f-01dfb84a35cd"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:48:17 crc kubenswrapper[4702]: I1125 13:48:17.131840 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cb1a228-426f-41cc-a15f-01dfb84a35cd-kube-api-access-jgbkn" (OuterVolumeSpecName: "kube-api-access-jgbkn") pod "8cb1a228-426f-41cc-a15f-01dfb84a35cd" (UID: "8cb1a228-426f-41cc-a15f-01dfb84a35cd"). InnerVolumeSpecName "kube-api-access-jgbkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:48:17 crc kubenswrapper[4702]: I1125 13:48:17.136339 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cb1a228-426f-41cc-a15f-01dfb84a35cd-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8cb1a228-426f-41cc-a15f-01dfb84a35cd" (UID: "8cb1a228-426f-41cc-a15f-01dfb84a35cd"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:48:17 crc kubenswrapper[4702]: I1125 13:48:17.225181 4702 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8cb1a228-426f-41cc-a15f-01dfb84a35cd-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 13:48:17 crc kubenswrapper[4702]: I1125 13:48:17.225245 4702 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8cb1a228-426f-41cc-a15f-01dfb84a35cd-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 13:48:17 crc kubenswrapper[4702]: I1125 13:48:17.225272 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgbkn\" (UniqueName: \"kubernetes.io/projected/8cb1a228-426f-41cc-a15f-01dfb84a35cd-kube-api-access-jgbkn\") on node \"crc\" DevicePath \"\"" Nov 25 13:48:17 crc kubenswrapper[4702]: I1125 13:48:17.359107 4702 patch_prober.go:28] interesting pod/router-default-5444994796-lmq49 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 13:48:17 crc kubenswrapper[4702]: [+]has-synced ok Nov 25 13:48:17 crc kubenswrapper[4702]: [+]process-running ok Nov 25 13:48:17 crc kubenswrapper[4702]: healthz check failed Nov 25 13:48:17 crc kubenswrapper[4702]: I1125 13:48:17.359674 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lmq49" podUID="6d7fc28a-71f2-415c-a99e-76e4ab7fbb70" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 13:48:17 crc kubenswrapper[4702]: I1125 13:48:17.662116 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 13:48:17 crc kubenswrapper[4702]: E1125 13:48:17.662828 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cb1a228-426f-41cc-a15f-01dfb84a35cd" containerName="collect-profiles" Nov 25 13:48:17 crc kubenswrapper[4702]: I1125 13:48:17.662943 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cb1a228-426f-41cc-a15f-01dfb84a35cd" containerName="collect-profiles" Nov 25 13:48:17 crc kubenswrapper[4702]: I1125 13:48:17.663140 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cb1a228-426f-41cc-a15f-01dfb84a35cd" containerName="collect-profiles" Nov 25 13:48:17 crc kubenswrapper[4702]: I1125 13:48:17.663820 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 13:48:17 crc kubenswrapper[4702]: I1125 13:48:17.673294 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 25 13:48:17 crc kubenswrapper[4702]: I1125 13:48:17.673929 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 25 13:48:17 crc kubenswrapper[4702]: I1125 13:48:17.682465 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 13:48:17 crc kubenswrapper[4702]: I1125 13:48:17.736058 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/99306a96-346a-4b84-9830-53e1c37ac665-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"99306a96-346a-4b84-9830-53e1c37ac665\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 13:48:17 crc kubenswrapper[4702]: I1125 13:48:17.736342 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/99306a96-346a-4b84-9830-53e1c37ac665-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"99306a96-346a-4b84-9830-53e1c37ac665\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 13:48:17 crc kubenswrapper[4702]: I1125 13:48:17.755199 4702 generic.go:334] "Generic (PLEG): container finished" podID="9ab03842-ddff-42ae-8378-8c5a6cc4fb35" containerID="3071ed6707bc47c77c110035c8e4a4546921639f6e2fd1b9b10fb0b28c0c8308" exitCode=0 Nov 25 13:48:17 crc kubenswrapper[4702]: I1125 13:48:17.755282 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"9ab03842-ddff-42ae-8378-8c5a6cc4fb35","Type":"ContainerDied","Data":"3071ed6707bc47c77c110035c8e4a4546921639f6e2fd1b9b10fb0b28c0c8308"} Nov 25 13:48:17 crc kubenswrapper[4702]: I1125 13:48:17.781111 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401305-j5rqf" event={"ID":"8cb1a228-426f-41cc-a15f-01dfb84a35cd","Type":"ContainerDied","Data":"78e85639211e12a8fdb3314315d00af8eb3d9b5c47eb1338468bb8d99c4842d8"} Nov 25 13:48:17 crc kubenswrapper[4702]: I1125 13:48:17.781174 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78e85639211e12a8fdb3314315d00af8eb3d9b5c47eb1338468bb8d99c4842d8" Nov 25 13:48:17 crc kubenswrapper[4702]: I1125 13:48:17.781543 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401305-j5rqf" Nov 25 13:48:17 crc kubenswrapper[4702]: I1125 13:48:17.838467 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/99306a96-346a-4b84-9830-53e1c37ac665-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"99306a96-346a-4b84-9830-53e1c37ac665\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 13:48:17 crc kubenswrapper[4702]: I1125 13:48:17.838554 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/99306a96-346a-4b84-9830-53e1c37ac665-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"99306a96-346a-4b84-9830-53e1c37ac665\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 13:48:17 crc kubenswrapper[4702]: I1125 13:48:17.838624 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/99306a96-346a-4b84-9830-53e1c37ac665-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"99306a96-346a-4b84-9830-53e1c37ac665\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 13:48:17 crc kubenswrapper[4702]: I1125 13:48:17.870630 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/99306a96-346a-4b84-9830-53e1c37ac665-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"99306a96-346a-4b84-9830-53e1c37ac665\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 13:48:18 crc kubenswrapper[4702]: I1125 13:48:18.010623 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 13:48:18 crc kubenswrapper[4702]: I1125 13:48:18.373312 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-lmq49" Nov 25 13:48:18 crc kubenswrapper[4702]: I1125 13:48:18.384276 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-lmq49" Nov 25 13:48:18 crc kubenswrapper[4702]: I1125 13:48:18.422072 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-jqmtw" Nov 25 13:48:18 crc kubenswrapper[4702]: I1125 13:48:18.855825 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 13:48:18 crc kubenswrapper[4702]: W1125 13:48:18.901027 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod99306a96_346a_4b84_9830_53e1c37ac665.slice/crio-eb311a0452a7d6ede453997fb7ab5ca36f537d7c04a06a3aa7787e6836e0b575 WatchSource:0}: Error finding container eb311a0452a7d6ede453997fb7ab5ca36f537d7c04a06a3aa7787e6836e0b575: Status 404 returned error can't find the container with id eb311a0452a7d6ede453997fb7ab5ca36f537d7c04a06a3aa7787e6836e0b575 Nov 25 13:48:19 crc kubenswrapper[4702]: I1125 13:48:19.161198 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 13:48:19 crc kubenswrapper[4702]: I1125 13:48:19.226904 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9ab03842-ddff-42ae-8378-8c5a6cc4fb35-kube-api-access\") pod \"9ab03842-ddff-42ae-8378-8c5a6cc4fb35\" (UID: \"9ab03842-ddff-42ae-8378-8c5a6cc4fb35\") " Nov 25 13:48:19 crc kubenswrapper[4702]: I1125 13:48:19.226962 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9ab03842-ddff-42ae-8378-8c5a6cc4fb35-kubelet-dir\") pod \"9ab03842-ddff-42ae-8378-8c5a6cc4fb35\" (UID: \"9ab03842-ddff-42ae-8378-8c5a6cc4fb35\") " Nov 25 13:48:19 crc kubenswrapper[4702]: I1125 13:48:19.227262 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ab03842-ddff-42ae-8378-8c5a6cc4fb35-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9ab03842-ddff-42ae-8378-8c5a6cc4fb35" (UID: "9ab03842-ddff-42ae-8378-8c5a6cc4fb35"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 13:48:19 crc kubenswrapper[4702]: I1125 13:48:19.252611 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ab03842-ddff-42ae-8378-8c5a6cc4fb35-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9ab03842-ddff-42ae-8378-8c5a6cc4fb35" (UID: "9ab03842-ddff-42ae-8378-8c5a6cc4fb35"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:48:19 crc kubenswrapper[4702]: I1125 13:48:19.328702 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9ab03842-ddff-42ae-8378-8c5a6cc4fb35-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 13:48:19 crc kubenswrapper[4702]: I1125 13:48:19.328742 4702 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9ab03842-ddff-42ae-8378-8c5a6cc4fb35-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 13:48:19 crc kubenswrapper[4702]: I1125 13:48:19.797313 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"9ab03842-ddff-42ae-8378-8c5a6cc4fb35","Type":"ContainerDied","Data":"17565ea5286d9300ba7085d138617f4441058f5e30e93087c018d7d80f6c2118"} Nov 25 13:48:19 crc kubenswrapper[4702]: I1125 13:48:19.797379 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17565ea5286d9300ba7085d138617f4441058f5e30e93087c018d7d80f6c2118" Nov 25 13:48:19 crc kubenswrapper[4702]: I1125 13:48:19.797464 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 13:48:19 crc kubenswrapper[4702]: I1125 13:48:19.823616 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"99306a96-346a-4b84-9830-53e1c37ac665","Type":"ContainerStarted","Data":"eb311a0452a7d6ede453997fb7ab5ca36f537d7c04a06a3aa7787e6836e0b575"} Nov 25 13:48:20 crc kubenswrapper[4702]: I1125 13:48:20.854540 4702 generic.go:334] "Generic (PLEG): container finished" podID="99306a96-346a-4b84-9830-53e1c37ac665" containerID="0ba57149ed2dab3e0622f8f04a7ba8d914a1005d880c68b5f14b82667968b57a" exitCode=0 Nov 25 13:48:20 crc kubenswrapper[4702]: I1125 13:48:20.854815 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"99306a96-346a-4b84-9830-53e1c37ac665","Type":"ContainerDied","Data":"0ba57149ed2dab3e0622f8f04a7ba8d914a1005d880c68b5f14b82667968b57a"} Nov 25 13:48:25 crc kubenswrapper[4702]: I1125 13:48:25.415878 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-mbmx2" Nov 25 13:48:25 crc kubenswrapper[4702]: I1125 13:48:25.422054 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-mbmx2" Nov 25 13:48:25 crc kubenswrapper[4702]: I1125 13:48:25.641201 4702 patch_prober.go:28] interesting pod/downloads-7954f5f757-8j9xk container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Nov 25 13:48:25 crc kubenswrapper[4702]: I1125 13:48:25.641365 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-8j9xk" podUID="ce097f6c-2894-4417-b66b-0f30d91f3889" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Nov 25 13:48:25 crc kubenswrapper[4702]: I1125 13:48:25.641795 4702 patch_prober.go:28] interesting pod/downloads-7954f5f757-8j9xk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Nov 25 13:48:25 crc kubenswrapper[4702]: I1125 13:48:25.641864 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8j9xk" podUID="ce097f6c-2894-4417-b66b-0f30d91f3889" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Nov 25 13:48:33 crc kubenswrapper[4702]: I1125 13:48:33.470968 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:48:35 crc kubenswrapper[4702]: I1125 13:48:35.659810 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-8j9xk" Nov 25 13:48:38 crc kubenswrapper[4702]: I1125 13:48:38.810832 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 13:48:38 crc kubenswrapper[4702]: I1125 13:48:38.954605 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/99306a96-346a-4b84-9830-53e1c37ac665-kube-api-access\") pod \"99306a96-346a-4b84-9830-53e1c37ac665\" (UID: \"99306a96-346a-4b84-9830-53e1c37ac665\") " Nov 25 13:48:38 crc kubenswrapper[4702]: I1125 13:48:38.954803 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/99306a96-346a-4b84-9830-53e1c37ac665-kubelet-dir\") pod \"99306a96-346a-4b84-9830-53e1c37ac665\" (UID: \"99306a96-346a-4b84-9830-53e1c37ac665\") " Nov 25 13:48:38 crc kubenswrapper[4702]: I1125 13:48:38.954999 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/99306a96-346a-4b84-9830-53e1c37ac665-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "99306a96-346a-4b84-9830-53e1c37ac665" (UID: "99306a96-346a-4b84-9830-53e1c37ac665"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 13:48:38 crc kubenswrapper[4702]: I1125 13:48:38.960825 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99306a96-346a-4b84-9830-53e1c37ac665-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "99306a96-346a-4b84-9830-53e1c37ac665" (UID: "99306a96-346a-4b84-9830-53e1c37ac665"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:48:38 crc kubenswrapper[4702]: I1125 13:48:38.995632 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"99306a96-346a-4b84-9830-53e1c37ac665","Type":"ContainerDied","Data":"eb311a0452a7d6ede453997fb7ab5ca36f537d7c04a06a3aa7787e6836e0b575"} Nov 25 13:48:38 crc kubenswrapper[4702]: I1125 13:48:38.995680 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb311a0452a7d6ede453997fb7ab5ca36f537d7c04a06a3aa7787e6836e0b575" Nov 25 13:48:38 crc kubenswrapper[4702]: I1125 13:48:38.995689 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 13:48:39 crc kubenswrapper[4702]: I1125 13:48:39.056858 4702 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/99306a96-346a-4b84-9830-53e1c37ac665-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 13:48:39 crc kubenswrapper[4702]: I1125 13:48:39.056905 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/99306a96-346a-4b84-9830-53e1c37ac665-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 13:48:46 crc kubenswrapper[4702]: I1125 13:48:46.372418 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4zsfb" Nov 25 13:48:52 crc kubenswrapper[4702]: E1125 13:48:52.025386 4702 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 25 13:48:52 crc kubenswrapper[4702]: E1125 13:48:52.026214 4702 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l5b4t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-pwlcj_openshift-marketplace(66c1bf9b-3a89-427b-9bba-009bdafcfcc8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 13:48:52 crc kubenswrapper[4702]: E1125 13:48:52.027894 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-pwlcj" podUID="66c1bf9b-3a89-427b-9bba-009bdafcfcc8" Nov 25 13:48:56 crc kubenswrapper[4702]: E1125 13:48:56.212352 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-pwlcj" podUID="66c1bf9b-3a89-427b-9bba-009bdafcfcc8" Nov 25 13:48:56 crc kubenswrapper[4702]: E1125 13:48:56.364785 4702 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 25 13:48:56 crc kubenswrapper[4702]: E1125 13:48:56.365318 4702 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-swzxn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-jq6gh_openshift-marketplace(f5b74e23-72c6-4278-9dc7-909f528d344b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 13:48:56 crc kubenswrapper[4702]: E1125 13:48:56.367917 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-jq6gh" podUID="f5b74e23-72c6-4278-9dc7-909f528d344b" Nov 25 13:48:56 crc kubenswrapper[4702]: E1125 13:48:56.430303 4702 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 25 13:48:56 crc kubenswrapper[4702]: E1125 13:48:56.430459 4702 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qr26k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-9bfw8_openshift-marketplace(3babd873-f61e-40d1-a42f-8a2f32f3ca40): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 13:48:56 crc kubenswrapper[4702]: E1125 13:48:56.431712 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-9bfw8" podUID="3babd873-f61e-40d1-a42f-8a2f32f3ca40" Nov 25 13:48:56 crc kubenswrapper[4702]: E1125 13:48:56.443174 4702 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 25 13:48:56 crc kubenswrapper[4702]: E1125 13:48:56.443397 4702 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4wncr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-rh58q_openshift-marketplace(75d44e5f-cb22-4c92-8695-b9b8128d8724): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 13:48:56 crc kubenswrapper[4702]: E1125 13:48:56.444596 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-rh58q" podUID="75d44e5f-cb22-4c92-8695-b9b8128d8724" Nov 25 13:48:57 crc kubenswrapper[4702]: I1125 13:48:57.536542 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 25 13:48:57 crc kubenswrapper[4702]: E1125 13:48:57.537295 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99306a96-346a-4b84-9830-53e1c37ac665" containerName="pruner" Nov 25 13:48:57 crc kubenswrapper[4702]: I1125 13:48:57.537310 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="99306a96-346a-4b84-9830-53e1c37ac665" containerName="pruner" Nov 25 13:48:57 crc kubenswrapper[4702]: E1125 13:48:57.537328 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ab03842-ddff-42ae-8378-8c5a6cc4fb35" containerName="pruner" Nov 25 13:48:57 crc kubenswrapper[4702]: I1125 13:48:57.537335 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ab03842-ddff-42ae-8378-8c5a6cc4fb35" containerName="pruner" Nov 25 13:48:57 crc kubenswrapper[4702]: I1125 13:48:57.542583 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ab03842-ddff-42ae-8378-8c5a6cc4fb35" containerName="pruner" Nov 25 13:48:57 crc kubenswrapper[4702]: I1125 13:48:57.542636 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="99306a96-346a-4b84-9830-53e1c37ac665" containerName="pruner" Nov 25 13:48:57 crc kubenswrapper[4702]: I1125 13:48:57.543223 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 13:48:57 crc kubenswrapper[4702]: I1125 13:48:57.545509 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 25 13:48:57 crc kubenswrapper[4702]: I1125 13:48:57.547189 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 25 13:48:57 crc kubenswrapper[4702]: I1125 13:48:57.547215 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 25 13:48:57 crc kubenswrapper[4702]: I1125 13:48:57.653222 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bd7f1a2b-31d5-4186-a698-a1d4c011a6f6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bd7f1a2b-31d5-4186-a698-a1d4c011a6f6\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 13:48:57 crc kubenswrapper[4702]: I1125 13:48:57.653426 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bd7f1a2b-31d5-4186-a698-a1d4c011a6f6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bd7f1a2b-31d5-4186-a698-a1d4c011a6f6\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 13:48:57 crc kubenswrapper[4702]: I1125 13:48:57.754915 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bd7f1a2b-31d5-4186-a698-a1d4c011a6f6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bd7f1a2b-31d5-4186-a698-a1d4c011a6f6\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 13:48:57 crc kubenswrapper[4702]: I1125 13:48:57.755002 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bd7f1a2b-31d5-4186-a698-a1d4c011a6f6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bd7f1a2b-31d5-4186-a698-a1d4c011a6f6\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 13:48:57 crc kubenswrapper[4702]: I1125 13:48:57.755155 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bd7f1a2b-31d5-4186-a698-a1d4c011a6f6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bd7f1a2b-31d5-4186-a698-a1d4c011a6f6\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 13:48:57 crc kubenswrapper[4702]: I1125 13:48:57.785025 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bd7f1a2b-31d5-4186-a698-a1d4c011a6f6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bd7f1a2b-31d5-4186-a698-a1d4c011a6f6\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 13:48:57 crc kubenswrapper[4702]: I1125 13:48:57.868366 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 13:48:57 crc kubenswrapper[4702]: E1125 13:48:57.880913 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-9bfw8" podUID="3babd873-f61e-40d1-a42f-8a2f32f3ca40" Nov 25 13:48:57 crc kubenswrapper[4702]: E1125 13:48:57.881300 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-rh58q" podUID="75d44e5f-cb22-4c92-8695-b9b8128d8724" Nov 25 13:48:57 crc kubenswrapper[4702]: E1125 13:48:57.882009 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-jq6gh" podUID="f5b74e23-72c6-4278-9dc7-909f528d344b" Nov 25 13:48:57 crc kubenswrapper[4702]: E1125 13:48:57.967276 4702 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 25 13:48:57 crc kubenswrapper[4702]: E1125 13:48:57.967555 4702 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6znlx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-5xb7k_openshift-marketplace(c48d725b-8394-44f2-90e4-7ba6576fbd77): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 13:48:57 crc kubenswrapper[4702]: E1125 13:48:57.968780 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-5xb7k" podUID="c48d725b-8394-44f2-90e4-7ba6576fbd77" Nov 25 13:48:57 crc kubenswrapper[4702]: E1125 13:48:57.990219 4702 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 25 13:48:57 crc kubenswrapper[4702]: E1125 13:48:57.991224 4702 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ntsgr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-v9skf_openshift-marketplace(597c2ecd-d536-4eec-bc02-89c798cbf25c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 13:48:57 crc kubenswrapper[4702]: E1125 13:48:57.992604 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-v9skf" podUID="597c2ecd-d536-4eec-bc02-89c798cbf25c" Nov 25 13:48:57 crc kubenswrapper[4702]: E1125 13:48:57.998422 4702 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 25 13:48:57 crc kubenswrapper[4702]: E1125 13:48:57.998658 4702 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6hhjf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-xm8tp_openshift-marketplace(a78d7c57-6155-429a-bbcf-168c3fa26e2e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 13:48:58 crc kubenswrapper[4702]: E1125 13:48:57.999841 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-xm8tp" podUID="a78d7c57-6155-429a-bbcf-168c3fa26e2e" Nov 25 13:48:58 crc kubenswrapper[4702]: E1125 13:48:58.126415 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-xm8tp" podUID="a78d7c57-6155-429a-bbcf-168c3fa26e2e" Nov 25 13:48:58 crc kubenswrapper[4702]: E1125 13:48:58.126563 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-5xb7k" podUID="c48d725b-8394-44f2-90e4-7ba6576fbd77" Nov 25 13:48:58 crc kubenswrapper[4702]: E1125 13:48:58.126616 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-v9skf" podUID="597c2ecd-d536-4eec-bc02-89c798cbf25c" Nov 25 13:48:58 crc kubenswrapper[4702]: I1125 13:48:58.341085 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 25 13:48:58 crc kubenswrapper[4702]: W1125 13:48:58.349446 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podbd7f1a2b_31d5_4186_a698_a1d4c011a6f6.slice/crio-a5d392836119e1a22b9bb38b2a0f88015bdb4dc733a4a65f4387ab1e20566aa0 WatchSource:0}: Error finding container a5d392836119e1a22b9bb38b2a0f88015bdb4dc733a4a65f4387ab1e20566aa0: Status 404 returned error can't find the container with id a5d392836119e1a22b9bb38b2a0f88015bdb4dc733a4a65f4387ab1e20566aa0 Nov 25 13:48:58 crc kubenswrapper[4702]: I1125 13:48:58.674036 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:48:58 crc kubenswrapper[4702]: I1125 13:48:58.674122 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:48:58 crc kubenswrapper[4702]: I1125 13:48:58.676640 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 25 13:48:58 crc kubenswrapper[4702]: I1125 13:48:58.678268 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 25 13:48:58 crc kubenswrapper[4702]: I1125 13:48:58.686038 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:48:58 crc kubenswrapper[4702]: I1125 13:48:58.691825 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:48:58 crc kubenswrapper[4702]: I1125 13:48:58.775776 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:48:58 crc kubenswrapper[4702]: I1125 13:48:58.775843 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:48:58 crc kubenswrapper[4702]: I1125 13:48:58.777860 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 25 13:48:58 crc kubenswrapper[4702]: I1125 13:48:58.788929 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 25 13:48:58 crc kubenswrapper[4702]: I1125 13:48:58.800192 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:48:58 crc kubenswrapper[4702]: I1125 13:48:58.800475 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:48:58 crc kubenswrapper[4702]: I1125 13:48:58.813953 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 13:48:58 crc kubenswrapper[4702]: I1125 13:48:58.824117 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:48:58 crc kubenswrapper[4702]: I1125 13:48:58.837486 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 13:48:59 crc kubenswrapper[4702]: I1125 13:48:59.132251 4702 generic.go:334] "Generic (PLEG): container finished" podID="68f8eda2-8222-4ecb-a2d0-316ed8398d58" containerID="5c8e883a9434d217bba8c7100e0a34f0c64b68b0a01d0f5520131ae3aaf78e12" exitCode=0 Nov 25 13:48:59 crc kubenswrapper[4702]: I1125 13:48:59.132464 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ppzc7" event={"ID":"68f8eda2-8222-4ecb-a2d0-316ed8398d58","Type":"ContainerDied","Data":"5c8e883a9434d217bba8c7100e0a34f0c64b68b0a01d0f5520131ae3aaf78e12"} Nov 25 13:48:59 crc kubenswrapper[4702]: I1125 13:48:59.135336 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"bd7f1a2b-31d5-4186-a698-a1d4c011a6f6","Type":"ContainerStarted","Data":"6186b8bde73b51cb876355f047abc4499a8a733bc50f51cd3c8201754dac7175"} Nov 25 13:48:59 crc kubenswrapper[4702]: I1125 13:48:59.135375 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"bd7f1a2b-31d5-4186-a698-a1d4c011a6f6","Type":"ContainerStarted","Data":"a5d392836119e1a22b9bb38b2a0f88015bdb4dc733a4a65f4387ab1e20566aa0"} Nov 25 13:48:59 crc kubenswrapper[4702]: I1125 13:48:59.241579 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=2.241546236 podStartE2EDuration="2.241546236s" podCreationTimestamp="2025-11-25 13:48:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:48:59.180020537 +0000 UTC m=+148.342670987" watchObservedRunningTime="2025-11-25 13:48:59.241546236 +0000 UTC m=+148.404196686" Nov 25 13:48:59 crc kubenswrapper[4702]: W1125 13:48:59.325986 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-14c4707860769f9c555969e85ac504d2a97f0002d649449cb247759c5ae21ed7 WatchSource:0}: Error finding container 14c4707860769f9c555969e85ac504d2a97f0002d649449cb247759c5ae21ed7: Status 404 returned error can't find the container with id 14c4707860769f9c555969e85ac504d2a97f0002d649449cb247759c5ae21ed7 Nov 25 13:48:59 crc kubenswrapper[4702]: W1125 13:48:59.343824 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-3929539ce5c12b86ddc3029527f6ee8ce6a73b930bdcbb4d55c6affe4542dfcd WatchSource:0}: Error finding container 3929539ce5c12b86ddc3029527f6ee8ce6a73b930bdcbb4d55c6affe4542dfcd: Status 404 returned error can't find the container with id 3929539ce5c12b86ddc3029527f6ee8ce6a73b930bdcbb4d55c6affe4542dfcd Nov 25 13:49:00 crc kubenswrapper[4702]: I1125 13:49:00.143935 4702 generic.go:334] "Generic (PLEG): container finished" podID="bd7f1a2b-31d5-4186-a698-a1d4c011a6f6" containerID="6186b8bde73b51cb876355f047abc4499a8a733bc50f51cd3c8201754dac7175" exitCode=0 Nov 25 13:49:00 crc kubenswrapper[4702]: I1125 13:49:00.144103 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"bd7f1a2b-31d5-4186-a698-a1d4c011a6f6","Type":"ContainerDied","Data":"6186b8bde73b51cb876355f047abc4499a8a733bc50f51cd3c8201754dac7175"} Nov 25 13:49:00 crc kubenswrapper[4702]: I1125 13:49:00.147949 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"3f8f9365d09817e44c09eb3164d25c10085fcda5544bea744ce0a0bb2bb2a42b"} Nov 25 13:49:00 crc kubenswrapper[4702]: I1125 13:49:00.147990 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"3929539ce5c12b86ddc3029527f6ee8ce6a73b930bdcbb4d55c6affe4542dfcd"} Nov 25 13:49:00 crc kubenswrapper[4702]: I1125 13:49:00.151021 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ppzc7" event={"ID":"68f8eda2-8222-4ecb-a2d0-316ed8398d58","Type":"ContainerStarted","Data":"93047c8e45a295c8afe15bafe26622ba661947c901b3e30712219f6388c35216"} Nov 25 13:49:00 crc kubenswrapper[4702]: I1125 13:49:00.153561 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"d679d4044af82877928f5422950ab2e61f11e554df4f944b55f2eec3ad13b271"} Nov 25 13:49:00 crc kubenswrapper[4702]: I1125 13:49:00.153600 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"14c4707860769f9c555969e85ac504d2a97f0002d649449cb247759c5ae21ed7"} Nov 25 13:49:00 crc kubenswrapper[4702]: I1125 13:49:00.153783 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:49:00 crc kubenswrapper[4702]: I1125 13:49:00.158528 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"0bc4fd8d63c89480306fd0eee58847683f1c24ab25b1084ca2cdb6d42864ba38"} Nov 25 13:49:00 crc kubenswrapper[4702]: I1125 13:49:00.158565 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"850646cf3521d3b15a00558a4f0be1452d2558bcb4fd974a7a8790f548c034e3"} Nov 25 13:49:00 crc kubenswrapper[4702]: I1125 13:49:00.269187 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ppzc7" podStartSLOduration=2.333231988 podStartE2EDuration="46.269163555s" podCreationTimestamp="2025-11-25 13:48:14 +0000 UTC" firstStartedPulling="2025-11-25 13:48:15.691173178 +0000 UTC m=+104.853823628" lastFinishedPulling="2025-11-25 13:48:59.627104745 +0000 UTC m=+148.789755195" observedRunningTime="2025-11-25 13:49:00.265408958 +0000 UTC m=+149.428059438" watchObservedRunningTime="2025-11-25 13:49:00.269163555 +0000 UTC m=+149.431814005" Nov 25 13:49:01 crc kubenswrapper[4702]: I1125 13:49:01.432908 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 13:49:01 crc kubenswrapper[4702]: I1125 13:49:01.522680 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bd7f1a2b-31d5-4186-a698-a1d4c011a6f6-kubelet-dir\") pod \"bd7f1a2b-31d5-4186-a698-a1d4c011a6f6\" (UID: \"bd7f1a2b-31d5-4186-a698-a1d4c011a6f6\") " Nov 25 13:49:01 crc kubenswrapper[4702]: I1125 13:49:01.522875 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bd7f1a2b-31d5-4186-a698-a1d4c011a6f6-kube-api-access\") pod \"bd7f1a2b-31d5-4186-a698-a1d4c011a6f6\" (UID: \"bd7f1a2b-31d5-4186-a698-a1d4c011a6f6\") " Nov 25 13:49:01 crc kubenswrapper[4702]: I1125 13:49:01.522873 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd7f1a2b-31d5-4186-a698-a1d4c011a6f6-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "bd7f1a2b-31d5-4186-a698-a1d4c011a6f6" (UID: "bd7f1a2b-31d5-4186-a698-a1d4c011a6f6"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 13:49:01 crc kubenswrapper[4702]: I1125 13:49:01.523223 4702 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bd7f1a2b-31d5-4186-a698-a1d4c011a6f6-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:01 crc kubenswrapper[4702]: I1125 13:49:01.531407 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd7f1a2b-31d5-4186-a698-a1d4c011a6f6-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "bd7f1a2b-31d5-4186-a698-a1d4c011a6f6" (UID: "bd7f1a2b-31d5-4186-a698-a1d4c011a6f6"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:49:01 crc kubenswrapper[4702]: I1125 13:49:01.624976 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bd7f1a2b-31d5-4186-a698-a1d4c011a6f6-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:02 crc kubenswrapper[4702]: I1125 13:49:02.172729 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"bd7f1a2b-31d5-4186-a698-a1d4c011a6f6","Type":"ContainerDied","Data":"a5d392836119e1a22b9bb38b2a0f88015bdb4dc733a4a65f4387ab1e20566aa0"} Nov 25 13:49:02 crc kubenswrapper[4702]: I1125 13:49:02.172777 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5d392836119e1a22b9bb38b2a0f88015bdb4dc733a4a65f4387ab1e20566aa0" Nov 25 13:49:02 crc kubenswrapper[4702]: I1125 13:49:02.172804 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 13:49:02 crc kubenswrapper[4702]: I1125 13:49:02.529522 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 25 13:49:02 crc kubenswrapper[4702]: E1125 13:49:02.530217 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd7f1a2b-31d5-4186-a698-a1d4c011a6f6" containerName="pruner" Nov 25 13:49:02 crc kubenswrapper[4702]: I1125 13:49:02.530336 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd7f1a2b-31d5-4186-a698-a1d4c011a6f6" containerName="pruner" Nov 25 13:49:02 crc kubenswrapper[4702]: I1125 13:49:02.530542 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd7f1a2b-31d5-4186-a698-a1d4c011a6f6" containerName="pruner" Nov 25 13:49:02 crc kubenswrapper[4702]: I1125 13:49:02.531093 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 13:49:02 crc kubenswrapper[4702]: I1125 13:49:02.540006 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 25 13:49:02 crc kubenswrapper[4702]: I1125 13:49:02.540306 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 25 13:49:02 crc kubenswrapper[4702]: I1125 13:49:02.546422 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 25 13:49:02 crc kubenswrapper[4702]: I1125 13:49:02.641279 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/70ee4d8e-0a03-4949-8c82-dc84b9ae5699-kube-api-access\") pod \"installer-9-crc\" (UID: \"70ee4d8e-0a03-4949-8c82-dc84b9ae5699\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 13:49:02 crc kubenswrapper[4702]: I1125 13:49:02.641399 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/70ee4d8e-0a03-4949-8c82-dc84b9ae5699-var-lock\") pod \"installer-9-crc\" (UID: \"70ee4d8e-0a03-4949-8c82-dc84b9ae5699\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 13:49:02 crc kubenswrapper[4702]: I1125 13:49:02.641435 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/70ee4d8e-0a03-4949-8c82-dc84b9ae5699-kubelet-dir\") pod \"installer-9-crc\" (UID: \"70ee4d8e-0a03-4949-8c82-dc84b9ae5699\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 13:49:02 crc kubenswrapper[4702]: I1125 13:49:02.742288 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/70ee4d8e-0a03-4949-8c82-dc84b9ae5699-var-lock\") pod \"installer-9-crc\" (UID: \"70ee4d8e-0a03-4949-8c82-dc84b9ae5699\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 13:49:02 crc kubenswrapper[4702]: I1125 13:49:02.742367 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/70ee4d8e-0a03-4949-8c82-dc84b9ae5699-kubelet-dir\") pod \"installer-9-crc\" (UID: \"70ee4d8e-0a03-4949-8c82-dc84b9ae5699\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 13:49:02 crc kubenswrapper[4702]: I1125 13:49:02.742410 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/70ee4d8e-0a03-4949-8c82-dc84b9ae5699-kube-api-access\") pod \"installer-9-crc\" (UID: \"70ee4d8e-0a03-4949-8c82-dc84b9ae5699\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 13:49:02 crc kubenswrapper[4702]: I1125 13:49:02.742461 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/70ee4d8e-0a03-4949-8c82-dc84b9ae5699-kubelet-dir\") pod \"installer-9-crc\" (UID: \"70ee4d8e-0a03-4949-8c82-dc84b9ae5699\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 13:49:02 crc kubenswrapper[4702]: I1125 13:49:02.742460 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/70ee4d8e-0a03-4949-8c82-dc84b9ae5699-var-lock\") pod \"installer-9-crc\" (UID: \"70ee4d8e-0a03-4949-8c82-dc84b9ae5699\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 13:49:02 crc kubenswrapper[4702]: I1125 13:49:02.765903 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/70ee4d8e-0a03-4949-8c82-dc84b9ae5699-kube-api-access\") pod \"installer-9-crc\" (UID: \"70ee4d8e-0a03-4949-8c82-dc84b9ae5699\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 13:49:02 crc kubenswrapper[4702]: I1125 13:49:02.873017 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 13:49:03 crc kubenswrapper[4702]: I1125 13:49:03.131073 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 25 13:49:03 crc kubenswrapper[4702]: I1125 13:49:03.181519 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"70ee4d8e-0a03-4949-8c82-dc84b9ae5699","Type":"ContainerStarted","Data":"46dbcf4fc8f469bf842092cc4170f339734716afa08095f60cac4f78e9be70d5"} Nov 25 13:49:03 crc kubenswrapper[4702]: I1125 13:49:03.963368 4702 patch_prober.go:28] interesting pod/machine-config-daemon-qqr76 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 13:49:03 crc kubenswrapper[4702]: I1125 13:49:03.963937 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 13:49:04 crc kubenswrapper[4702]: I1125 13:49:04.188248 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"70ee4d8e-0a03-4949-8c82-dc84b9ae5699","Type":"ContainerStarted","Data":"39fec14b7d5465bc6e100123bdbeb78d28942068e22897c6aed71d85dc6af613"} Nov 25 13:49:04 crc kubenswrapper[4702]: I1125 13:49:04.912615 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ppzc7" Nov 25 13:49:04 crc kubenswrapper[4702]: I1125 13:49:04.913021 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ppzc7" Nov 25 13:49:06 crc kubenswrapper[4702]: I1125 13:49:06.049650 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ppzc7" podUID="68f8eda2-8222-4ecb-a2d0-316ed8398d58" containerName="registry-server" probeResult="failure" output=< Nov 25 13:49:06 crc kubenswrapper[4702]: timeout: failed to connect service ":50051" within 1s Nov 25 13:49:06 crc kubenswrapper[4702]: > Nov 25 13:49:08 crc kubenswrapper[4702]: I1125 13:49:08.821202 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=6.821183451 podStartE2EDuration="6.821183451s" podCreationTimestamp="2025-11-25 13:49:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:49:04.232353386 +0000 UTC m=+153.395003856" watchObservedRunningTime="2025-11-25 13:49:08.821183451 +0000 UTC m=+157.983833901" Nov 25 13:49:13 crc kubenswrapper[4702]: E1125 13:49:13.569778 4702 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3babd873_f61e_40d1_a42f_8a2f32f3ca40.slice/crio-de53627fae115bb4b90a8c7547462aba80a0eba2ef2c5c8997fa8a5363c9b99e.scope\": RecentStats: unable to find data in memory cache]" Nov 25 13:49:14 crc kubenswrapper[4702]: I1125 13:49:14.263365 4702 generic.go:334] "Generic (PLEG): container finished" podID="3babd873-f61e-40d1-a42f-8a2f32f3ca40" containerID="de53627fae115bb4b90a8c7547462aba80a0eba2ef2c5c8997fa8a5363c9b99e" exitCode=0 Nov 25 13:49:14 crc kubenswrapper[4702]: I1125 13:49:14.263457 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9bfw8" event={"ID":"3babd873-f61e-40d1-a42f-8a2f32f3ca40","Type":"ContainerDied","Data":"de53627fae115bb4b90a8c7547462aba80a0eba2ef2c5c8997fa8a5363c9b99e"} Nov 25 13:49:14 crc kubenswrapper[4702]: I1125 13:49:14.267381 4702 generic.go:334] "Generic (PLEG): container finished" podID="597c2ecd-d536-4eec-bc02-89c798cbf25c" containerID="899670a015b79035a701d6929e4b0dc3fb601ca8595069cc726ec84fc0befc9a" exitCode=0 Nov 25 13:49:14 crc kubenswrapper[4702]: I1125 13:49:14.267503 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v9skf" event={"ID":"597c2ecd-d536-4eec-bc02-89c798cbf25c","Type":"ContainerDied","Data":"899670a015b79035a701d6929e4b0dc3fb601ca8595069cc726ec84fc0befc9a"} Nov 25 13:49:15 crc kubenswrapper[4702]: I1125 13:49:15.002168 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ppzc7" Nov 25 13:49:15 crc kubenswrapper[4702]: I1125 13:49:15.046773 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ppzc7" Nov 25 13:49:15 crc kubenswrapper[4702]: I1125 13:49:15.276049 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rh58q" event={"ID":"75d44e5f-cb22-4c92-8695-b9b8128d8724","Type":"ContainerStarted","Data":"41fbbcc18094cd39841affd6f742d142bb78f167c84e731525d5f82fbfb9c7d5"} Nov 25 13:49:15 crc kubenswrapper[4702]: I1125 13:49:15.278220 4702 generic.go:334] "Generic (PLEG): container finished" podID="a78d7c57-6155-429a-bbcf-168c3fa26e2e" containerID="f84d2deb0ab6842ea62301968e93476f9b55eaa4beb47460111c184ad46e7c7d" exitCode=0 Nov 25 13:49:15 crc kubenswrapper[4702]: I1125 13:49:15.278292 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xm8tp" event={"ID":"a78d7c57-6155-429a-bbcf-168c3fa26e2e","Type":"ContainerDied","Data":"f84d2deb0ab6842ea62301968e93476f9b55eaa4beb47460111c184ad46e7c7d"} Nov 25 13:49:15 crc kubenswrapper[4702]: I1125 13:49:15.280781 4702 generic.go:334] "Generic (PLEG): container finished" podID="66c1bf9b-3a89-427b-9bba-009bdafcfcc8" containerID="324b02bfbcc635df3fc842caec60c5d6a61be5e70b0eeecff1af40baa8a1662f" exitCode=0 Nov 25 13:49:15 crc kubenswrapper[4702]: I1125 13:49:15.280840 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pwlcj" event={"ID":"66c1bf9b-3a89-427b-9bba-009bdafcfcc8","Type":"ContainerDied","Data":"324b02bfbcc635df3fc842caec60c5d6a61be5e70b0eeecff1af40baa8a1662f"} Nov 25 13:49:15 crc kubenswrapper[4702]: I1125 13:49:15.289105 4702 generic.go:334] "Generic (PLEG): container finished" podID="f5b74e23-72c6-4278-9dc7-909f528d344b" containerID="92fdee327707bb0c612cfd0011379e14b1ff8ee7af7c4eebc8db3c471a4eb0b9" exitCode=0 Nov 25 13:49:15 crc kubenswrapper[4702]: I1125 13:49:15.289159 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jq6gh" event={"ID":"f5b74e23-72c6-4278-9dc7-909f528d344b","Type":"ContainerDied","Data":"92fdee327707bb0c612cfd0011379e14b1ff8ee7af7c4eebc8db3c471a4eb0b9"} Nov 25 13:49:15 crc kubenswrapper[4702]: I1125 13:49:15.292457 4702 generic.go:334] "Generic (PLEG): container finished" podID="c48d725b-8394-44f2-90e4-7ba6576fbd77" containerID="bb703016b3f3e7ba0ed1fc7b5fbe384ad810e8504fab9d6ad9fe599070021619" exitCode=0 Nov 25 13:49:15 crc kubenswrapper[4702]: I1125 13:49:15.293112 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5xb7k" event={"ID":"c48d725b-8394-44f2-90e4-7ba6576fbd77","Type":"ContainerDied","Data":"bb703016b3f3e7ba0ed1fc7b5fbe384ad810e8504fab9d6ad9fe599070021619"} Nov 25 13:49:16 crc kubenswrapper[4702]: I1125 13:49:16.299289 4702 generic.go:334] "Generic (PLEG): container finished" podID="75d44e5f-cb22-4c92-8695-b9b8128d8724" containerID="41fbbcc18094cd39841affd6f742d142bb78f167c84e731525d5f82fbfb9c7d5" exitCode=0 Nov 25 13:49:16 crc kubenswrapper[4702]: I1125 13:49:16.299329 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rh58q" event={"ID":"75d44e5f-cb22-4c92-8695-b9b8128d8724","Type":"ContainerDied","Data":"41fbbcc18094cd39841affd6f742d142bb78f167c84e731525d5f82fbfb9c7d5"} Nov 25 13:49:17 crc kubenswrapper[4702]: I1125 13:49:17.308303 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9bfw8" event={"ID":"3babd873-f61e-40d1-a42f-8a2f32f3ca40","Type":"ContainerStarted","Data":"70a3c21ba87e7f31241fc6f75f335ca1037f44f58869338bce42afca9628e267"} Nov 25 13:49:17 crc kubenswrapper[4702]: I1125 13:49:17.310669 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v9skf" event={"ID":"597c2ecd-d536-4eec-bc02-89c798cbf25c","Type":"ContainerStarted","Data":"de2131333c95abcea4d33eff4ac4f0c9854933323f5b386caee89bf5fe159ee3"} Nov 25 13:49:17 crc kubenswrapper[4702]: I1125 13:49:17.312930 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jq6gh" event={"ID":"f5b74e23-72c6-4278-9dc7-909f528d344b","Type":"ContainerStarted","Data":"9aeb281f96a9e4909ba74fcb0246b38a51b1578cf75be09905077750c19b1eb1"} Nov 25 13:49:17 crc kubenswrapper[4702]: I1125 13:49:17.358348 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-v9skf" podStartSLOduration=3.170329185 podStartE2EDuration="1m6.358326535s" podCreationTimestamp="2025-11-25 13:48:11 +0000 UTC" firstStartedPulling="2025-11-25 13:48:13.593046226 +0000 UTC m=+102.755696676" lastFinishedPulling="2025-11-25 13:49:16.781043576 +0000 UTC m=+165.943694026" observedRunningTime="2025-11-25 13:49:17.356178893 +0000 UTC m=+166.518829343" watchObservedRunningTime="2025-11-25 13:49:17.358326535 +0000 UTC m=+166.520976985" Nov 25 13:49:17 crc kubenswrapper[4702]: I1125 13:49:17.359600 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9bfw8" podStartSLOduration=4.933607204 podStartE2EDuration="1m6.359589961s" podCreationTimestamp="2025-11-25 13:48:11 +0000 UTC" firstStartedPulling="2025-11-25 13:48:14.608844618 +0000 UTC m=+103.771495058" lastFinishedPulling="2025-11-25 13:49:16.034827365 +0000 UTC m=+165.197477815" observedRunningTime="2025-11-25 13:49:17.335767734 +0000 UTC m=+166.498418194" watchObservedRunningTime="2025-11-25 13:49:17.359589961 +0000 UTC m=+166.522240411" Nov 25 13:49:17 crc kubenswrapper[4702]: I1125 13:49:17.379763 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jq6gh" podStartSLOduration=3.200529748 podStartE2EDuration="1m4.379740573s" podCreationTimestamp="2025-11-25 13:48:13 +0000 UTC" firstStartedPulling="2025-11-25 13:48:15.638359419 +0000 UTC m=+104.801009869" lastFinishedPulling="2025-11-25 13:49:16.817570244 +0000 UTC m=+165.980220694" observedRunningTime="2025-11-25 13:49:17.377394416 +0000 UTC m=+166.540044866" watchObservedRunningTime="2025-11-25 13:49:17.379740573 +0000 UTC m=+166.542391023" Nov 25 13:49:18 crc kubenswrapper[4702]: I1125 13:49:18.321020 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pwlcj" event={"ID":"66c1bf9b-3a89-427b-9bba-009bdafcfcc8","Type":"ContainerStarted","Data":"a8f0d3317038197b407ec5e59b13dc2e6109556f4b200e95589639b0b6630fd1"} Nov 25 13:49:19 crc kubenswrapper[4702]: I1125 13:49:19.328755 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5xb7k" event={"ID":"c48d725b-8394-44f2-90e4-7ba6576fbd77","Type":"ContainerStarted","Data":"9928413bfabd1ba62692153a1e5bd900b6119e5be7254990ff92e93effa6113a"} Nov 25 13:49:19 crc kubenswrapper[4702]: I1125 13:49:19.331069 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rh58q" event={"ID":"75d44e5f-cb22-4c92-8695-b9b8128d8724","Type":"ContainerStarted","Data":"933261426a8ef97c9c22eac32e69ef98a98c29682be91f129fc80bbcce1ab3c2"} Nov 25 13:49:19 crc kubenswrapper[4702]: I1125 13:49:19.333120 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xm8tp" event={"ID":"a78d7c57-6155-429a-bbcf-168c3fa26e2e","Type":"ContainerStarted","Data":"d39bac80e652f67ba6b8913ac9af87cd406e8dae75b80d389d0a009862d3296c"} Nov 25 13:49:19 crc kubenswrapper[4702]: I1125 13:49:19.855298 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pwlcj" podStartSLOduration=5.827919293 podStartE2EDuration="1m8.855253438s" podCreationTimestamp="2025-11-25 13:48:11 +0000 UTC" firstStartedPulling="2025-11-25 13:48:14.622455174 +0000 UTC m=+103.785105624" lastFinishedPulling="2025-11-25 13:49:17.649789319 +0000 UTC m=+166.812439769" observedRunningTime="2025-11-25 13:49:18.338707531 +0000 UTC m=+167.501357981" watchObservedRunningTime="2025-11-25 13:49:19.855253438 +0000 UTC m=+169.017903888" Nov 25 13:49:19 crc kubenswrapper[4702]: I1125 13:49:19.860045 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5xb7k" podStartSLOduration=5.201569726 podStartE2EDuration="1m8.860020844s" podCreationTimestamp="2025-11-25 13:48:11 +0000 UTC" firstStartedPulling="2025-11-25 13:48:14.607515826 +0000 UTC m=+103.770166276" lastFinishedPulling="2025-11-25 13:49:18.265966954 +0000 UTC m=+167.428617394" observedRunningTime="2025-11-25 13:49:19.8380799 +0000 UTC m=+169.000730370" watchObservedRunningTime="2025-11-25 13:49:19.860020844 +0000 UTC m=+169.022671294" Nov 25 13:49:19 crc kubenswrapper[4702]: I1125 13:49:19.867698 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xm8tp" podStartSLOduration=4.615565756 podStartE2EDuration="1m6.867680192s" podCreationTimestamp="2025-11-25 13:48:13 +0000 UTC" firstStartedPulling="2025-11-25 13:48:15.678768173 +0000 UTC m=+104.841418623" lastFinishedPulling="2025-11-25 13:49:17.930882609 +0000 UTC m=+167.093533059" observedRunningTime="2025-11-25 13:49:19.862596817 +0000 UTC m=+169.025247307" watchObservedRunningTime="2025-11-25 13:49:19.867680192 +0000 UTC m=+169.030330642" Nov 25 13:49:19 crc kubenswrapper[4702]: I1125 13:49:19.889428 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rh58q" podStartSLOduration=4.238539989 podStartE2EDuration="1m5.88940746s" podCreationTimestamp="2025-11-25 13:48:14 +0000 UTC" firstStartedPulling="2025-11-25 13:48:16.692835803 +0000 UTC m=+105.855486253" lastFinishedPulling="2025-11-25 13:49:18.343703274 +0000 UTC m=+167.506353724" observedRunningTime="2025-11-25 13:49:19.88659803 +0000 UTC m=+169.049248500" watchObservedRunningTime="2025-11-25 13:49:19.88940746 +0000 UTC m=+169.052057910" Nov 25 13:49:21 crc kubenswrapper[4702]: I1125 13:49:21.883827 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-v9skf" Nov 25 13:49:21 crc kubenswrapper[4702]: I1125 13:49:21.883892 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-v9skf" Nov 25 13:49:21 crc kubenswrapper[4702]: I1125 13:49:21.923264 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-v9skf" Nov 25 13:49:21 crc kubenswrapper[4702]: I1125 13:49:21.970111 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pwlcj" Nov 25 13:49:21 crc kubenswrapper[4702]: I1125 13:49:21.970182 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pwlcj" Nov 25 13:49:22 crc kubenswrapper[4702]: I1125 13:49:22.016180 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pwlcj" Nov 25 13:49:22 crc kubenswrapper[4702]: I1125 13:49:22.152847 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5xb7k" Nov 25 13:49:22 crc kubenswrapper[4702]: I1125 13:49:22.153293 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5xb7k" Nov 25 13:49:22 crc kubenswrapper[4702]: I1125 13:49:22.208676 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5xb7k" Nov 25 13:49:22 crc kubenswrapper[4702]: I1125 13:49:22.329504 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9bfw8" Nov 25 13:49:22 crc kubenswrapper[4702]: I1125 13:49:22.329781 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9bfw8" Nov 25 13:49:22 crc kubenswrapper[4702]: I1125 13:49:22.376020 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9bfw8" Nov 25 13:49:22 crc kubenswrapper[4702]: I1125 13:49:22.388722 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-v9skf" Nov 25 13:49:22 crc kubenswrapper[4702]: I1125 13:49:22.390223 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pwlcj" Nov 25 13:49:22 crc kubenswrapper[4702]: I1125 13:49:22.421521 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9bfw8" Nov 25 13:49:23 crc kubenswrapper[4702]: I1125 13:49:23.952992 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jq6gh" Nov 25 13:49:23 crc kubenswrapper[4702]: I1125 13:49:23.953057 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jq6gh" Nov 25 13:49:23 crc kubenswrapper[4702]: I1125 13:49:23.998599 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jq6gh" Nov 25 13:49:24 crc kubenswrapper[4702]: I1125 13:49:24.035687 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9bfw8"] Nov 25 13:49:24 crc kubenswrapper[4702]: I1125 13:49:24.329076 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xm8tp" Nov 25 13:49:24 crc kubenswrapper[4702]: I1125 13:49:24.330178 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xm8tp" Nov 25 13:49:24 crc kubenswrapper[4702]: I1125 13:49:24.360738 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9bfw8" podUID="3babd873-f61e-40d1-a42f-8a2f32f3ca40" containerName="registry-server" containerID="cri-o://70a3c21ba87e7f31241fc6f75f335ca1037f44f58869338bce42afca9628e267" gracePeriod=2 Nov 25 13:49:24 crc kubenswrapper[4702]: I1125 13:49:24.385722 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xm8tp" Nov 25 13:49:24 crc kubenswrapper[4702]: I1125 13:49:24.408644 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jq6gh" Nov 25 13:49:25 crc kubenswrapper[4702]: I1125 13:49:25.322331 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rh58q" Nov 25 13:49:25 crc kubenswrapper[4702]: I1125 13:49:25.323751 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rh58q" Nov 25 13:49:25 crc kubenswrapper[4702]: I1125 13:49:25.377474 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rh58q" Nov 25 13:49:25 crc kubenswrapper[4702]: I1125 13:49:25.431737 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xm8tp" Nov 25 13:49:25 crc kubenswrapper[4702]: I1125 13:49:25.435144 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rh58q" Nov 25 13:49:26 crc kubenswrapper[4702]: I1125 13:49:26.234975 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xm8tp"] Nov 25 13:49:26 crc kubenswrapper[4702]: I1125 13:49:26.375634 4702 generic.go:334] "Generic (PLEG): container finished" podID="3babd873-f61e-40d1-a42f-8a2f32f3ca40" containerID="70a3c21ba87e7f31241fc6f75f335ca1037f44f58869338bce42afca9628e267" exitCode=0 Nov 25 13:49:26 crc kubenswrapper[4702]: I1125 13:49:26.375727 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9bfw8" event={"ID":"3babd873-f61e-40d1-a42f-8a2f32f3ca40","Type":"ContainerDied","Data":"70a3c21ba87e7f31241fc6f75f335ca1037f44f58869338bce42afca9628e267"} Nov 25 13:49:27 crc kubenswrapper[4702]: I1125 13:49:27.011431 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9bfw8" Nov 25 13:49:27 crc kubenswrapper[4702]: I1125 13:49:27.104823 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qr26k\" (UniqueName: \"kubernetes.io/projected/3babd873-f61e-40d1-a42f-8a2f32f3ca40-kube-api-access-qr26k\") pod \"3babd873-f61e-40d1-a42f-8a2f32f3ca40\" (UID: \"3babd873-f61e-40d1-a42f-8a2f32f3ca40\") " Nov 25 13:49:27 crc kubenswrapper[4702]: I1125 13:49:27.104893 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3babd873-f61e-40d1-a42f-8a2f32f3ca40-catalog-content\") pod \"3babd873-f61e-40d1-a42f-8a2f32f3ca40\" (UID: \"3babd873-f61e-40d1-a42f-8a2f32f3ca40\") " Nov 25 13:49:27 crc kubenswrapper[4702]: I1125 13:49:27.104995 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3babd873-f61e-40d1-a42f-8a2f32f3ca40-utilities\") pod \"3babd873-f61e-40d1-a42f-8a2f32f3ca40\" (UID: \"3babd873-f61e-40d1-a42f-8a2f32f3ca40\") " Nov 25 13:49:27 crc kubenswrapper[4702]: I1125 13:49:27.105923 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3babd873-f61e-40d1-a42f-8a2f32f3ca40-utilities" (OuterVolumeSpecName: "utilities") pod "3babd873-f61e-40d1-a42f-8a2f32f3ca40" (UID: "3babd873-f61e-40d1-a42f-8a2f32f3ca40"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:49:27 crc kubenswrapper[4702]: I1125 13:49:27.112189 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3babd873-f61e-40d1-a42f-8a2f32f3ca40-kube-api-access-qr26k" (OuterVolumeSpecName: "kube-api-access-qr26k") pod "3babd873-f61e-40d1-a42f-8a2f32f3ca40" (UID: "3babd873-f61e-40d1-a42f-8a2f32f3ca40"). InnerVolumeSpecName "kube-api-access-qr26k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:49:27 crc kubenswrapper[4702]: I1125 13:49:27.206709 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3babd873-f61e-40d1-a42f-8a2f32f3ca40-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:27 crc kubenswrapper[4702]: I1125 13:49:27.206751 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qr26k\" (UniqueName: \"kubernetes.io/projected/3babd873-f61e-40d1-a42f-8a2f32f3ca40-kube-api-access-qr26k\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:27 crc kubenswrapper[4702]: I1125 13:49:27.385079 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9bfw8" event={"ID":"3babd873-f61e-40d1-a42f-8a2f32f3ca40","Type":"ContainerDied","Data":"cd3f34747ba3c2c05a0acdea6cd6f1f5f8db42f2bdd2a5055b3110d74ba9deb1"} Nov 25 13:49:27 crc kubenswrapper[4702]: I1125 13:49:27.385145 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9bfw8" Nov 25 13:49:27 crc kubenswrapper[4702]: I1125 13:49:27.385203 4702 scope.go:117] "RemoveContainer" containerID="70a3c21ba87e7f31241fc6f75f335ca1037f44f58869338bce42afca9628e267" Nov 25 13:49:27 crc kubenswrapper[4702]: I1125 13:49:27.385411 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xm8tp" podUID="a78d7c57-6155-429a-bbcf-168c3fa26e2e" containerName="registry-server" containerID="cri-o://d39bac80e652f67ba6b8913ac9af87cd406e8dae75b80d389d0a009862d3296c" gracePeriod=2 Nov 25 13:49:27 crc kubenswrapper[4702]: I1125 13:49:27.481298 4702 scope.go:117] "RemoveContainer" containerID="de53627fae115bb4b90a8c7547462aba80a0eba2ef2c5c8997fa8a5363c9b99e" Nov 25 13:49:27 crc kubenswrapper[4702]: I1125 13:49:27.511841 4702 scope.go:117] "RemoveContainer" containerID="996fb999cb3152d20808978aa1e2361941f291157495c8f7203d464c24ea18ca" Nov 25 13:49:28 crc kubenswrapper[4702]: I1125 13:49:28.438615 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rh58q"] Nov 25 13:49:28 crc kubenswrapper[4702]: I1125 13:49:28.438985 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rh58q" podUID="75d44e5f-cb22-4c92-8695-b9b8128d8724" containerName="registry-server" containerID="cri-o://933261426a8ef97c9c22eac32e69ef98a98c29682be91f129fc80bbcce1ab3c2" gracePeriod=2 Nov 25 13:49:29 crc kubenswrapper[4702]: I1125 13:49:29.401198 4702 generic.go:334] "Generic (PLEG): container finished" podID="a78d7c57-6155-429a-bbcf-168c3fa26e2e" containerID="d39bac80e652f67ba6b8913ac9af87cd406e8dae75b80d389d0a009862d3296c" exitCode=0 Nov 25 13:49:29 crc kubenswrapper[4702]: I1125 13:49:29.401316 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xm8tp" event={"ID":"a78d7c57-6155-429a-bbcf-168c3fa26e2e","Type":"ContainerDied","Data":"d39bac80e652f67ba6b8913ac9af87cd406e8dae75b80d389d0a009862d3296c"} Nov 25 13:49:29 crc kubenswrapper[4702]: I1125 13:49:29.896522 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-dfr9q"] Nov 25 13:49:30 crc kubenswrapper[4702]: I1125 13:49:30.044526 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xm8tp" Nov 25 13:49:30 crc kubenswrapper[4702]: I1125 13:49:30.146530 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hhjf\" (UniqueName: \"kubernetes.io/projected/a78d7c57-6155-429a-bbcf-168c3fa26e2e-kube-api-access-6hhjf\") pod \"a78d7c57-6155-429a-bbcf-168c3fa26e2e\" (UID: \"a78d7c57-6155-429a-bbcf-168c3fa26e2e\") " Nov 25 13:49:30 crc kubenswrapper[4702]: I1125 13:49:30.146608 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a78d7c57-6155-429a-bbcf-168c3fa26e2e-catalog-content\") pod \"a78d7c57-6155-429a-bbcf-168c3fa26e2e\" (UID: \"a78d7c57-6155-429a-bbcf-168c3fa26e2e\") " Nov 25 13:49:30 crc kubenswrapper[4702]: I1125 13:49:30.146655 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a78d7c57-6155-429a-bbcf-168c3fa26e2e-utilities\") pod \"a78d7c57-6155-429a-bbcf-168c3fa26e2e\" (UID: \"a78d7c57-6155-429a-bbcf-168c3fa26e2e\") " Nov 25 13:49:30 crc kubenswrapper[4702]: I1125 13:49:30.147557 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a78d7c57-6155-429a-bbcf-168c3fa26e2e-utilities" (OuterVolumeSpecName: "utilities") pod "a78d7c57-6155-429a-bbcf-168c3fa26e2e" (UID: "a78d7c57-6155-429a-bbcf-168c3fa26e2e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:49:30 crc kubenswrapper[4702]: I1125 13:49:30.163801 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a78d7c57-6155-429a-bbcf-168c3fa26e2e-kube-api-access-6hhjf" (OuterVolumeSpecName: "kube-api-access-6hhjf") pod "a78d7c57-6155-429a-bbcf-168c3fa26e2e" (UID: "a78d7c57-6155-429a-bbcf-168c3fa26e2e"). InnerVolumeSpecName "kube-api-access-6hhjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:49:30 crc kubenswrapper[4702]: I1125 13:49:30.167353 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a78d7c57-6155-429a-bbcf-168c3fa26e2e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a78d7c57-6155-429a-bbcf-168c3fa26e2e" (UID: "a78d7c57-6155-429a-bbcf-168c3fa26e2e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:49:30 crc kubenswrapper[4702]: I1125 13:49:30.248412 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hhjf\" (UniqueName: \"kubernetes.io/projected/a78d7c57-6155-429a-bbcf-168c3fa26e2e-kube-api-access-6hhjf\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:30 crc kubenswrapper[4702]: I1125 13:49:30.248450 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a78d7c57-6155-429a-bbcf-168c3fa26e2e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:30 crc kubenswrapper[4702]: I1125 13:49:30.248460 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a78d7c57-6155-429a-bbcf-168c3fa26e2e-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:30 crc kubenswrapper[4702]: I1125 13:49:30.409319 4702 generic.go:334] "Generic (PLEG): container finished" podID="75d44e5f-cb22-4c92-8695-b9b8128d8724" containerID="933261426a8ef97c9c22eac32e69ef98a98c29682be91f129fc80bbcce1ab3c2" exitCode=0 Nov 25 13:49:30 crc kubenswrapper[4702]: I1125 13:49:30.409412 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rh58q" event={"ID":"75d44e5f-cb22-4c92-8695-b9b8128d8724","Type":"ContainerDied","Data":"933261426a8ef97c9c22eac32e69ef98a98c29682be91f129fc80bbcce1ab3c2"} Nov 25 13:49:30 crc kubenswrapper[4702]: I1125 13:49:30.411986 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xm8tp" event={"ID":"a78d7c57-6155-429a-bbcf-168c3fa26e2e","Type":"ContainerDied","Data":"8ea4ebe9c214764646b49ce4213144f6b300e8a3ee5ba3819c2c0594266ba648"} Nov 25 13:49:30 crc kubenswrapper[4702]: I1125 13:49:30.412074 4702 scope.go:117] "RemoveContainer" containerID="d39bac80e652f67ba6b8913ac9af87cd406e8dae75b80d389d0a009862d3296c" Nov 25 13:49:30 crc kubenswrapper[4702]: I1125 13:49:30.412087 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xm8tp" Nov 25 13:49:30 crc kubenswrapper[4702]: I1125 13:49:30.436938 4702 scope.go:117] "RemoveContainer" containerID="f84d2deb0ab6842ea62301968e93476f9b55eaa4beb47460111c184ad46e7c7d" Nov 25 13:49:30 crc kubenswrapper[4702]: I1125 13:49:30.442573 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xm8tp"] Nov 25 13:49:30 crc kubenswrapper[4702]: I1125 13:49:30.445734 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xm8tp"] Nov 25 13:49:30 crc kubenswrapper[4702]: I1125 13:49:30.454911 4702 scope.go:117] "RemoveContainer" containerID="103f8ed0fc0c0650810aee82d33999bee26212a92d3a64ba13b61a471ceed5ed" Nov 25 13:49:31 crc kubenswrapper[4702]: I1125 13:49:31.313327 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3babd873-f61e-40d1-a42f-8a2f32f3ca40-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3babd873-f61e-40d1-a42f-8a2f32f3ca40" (UID: "3babd873-f61e-40d1-a42f-8a2f32f3ca40"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:49:31 crc kubenswrapper[4702]: I1125 13:49:31.362207 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3babd873-f61e-40d1-a42f-8a2f32f3ca40-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:31 crc kubenswrapper[4702]: I1125 13:49:31.557972 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rh58q" Nov 25 13:49:31 crc kubenswrapper[4702]: I1125 13:49:31.615465 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9bfw8"] Nov 25 13:49:31 crc kubenswrapper[4702]: I1125 13:49:31.619017 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9bfw8"] Nov 25 13:49:31 crc kubenswrapper[4702]: I1125 13:49:31.665278 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75d44e5f-cb22-4c92-8695-b9b8128d8724-utilities\") pod \"75d44e5f-cb22-4c92-8695-b9b8128d8724\" (UID: \"75d44e5f-cb22-4c92-8695-b9b8128d8724\") " Nov 25 13:49:31 crc kubenswrapper[4702]: I1125 13:49:31.665357 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75d44e5f-cb22-4c92-8695-b9b8128d8724-catalog-content\") pod \"75d44e5f-cb22-4c92-8695-b9b8128d8724\" (UID: \"75d44e5f-cb22-4c92-8695-b9b8128d8724\") " Nov 25 13:49:31 crc kubenswrapper[4702]: I1125 13:49:31.665394 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wncr\" (UniqueName: \"kubernetes.io/projected/75d44e5f-cb22-4c92-8695-b9b8128d8724-kube-api-access-4wncr\") pod \"75d44e5f-cb22-4c92-8695-b9b8128d8724\" (UID: \"75d44e5f-cb22-4c92-8695-b9b8128d8724\") " Nov 25 13:49:31 crc kubenswrapper[4702]: I1125 13:49:31.666000 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75d44e5f-cb22-4c92-8695-b9b8128d8724-utilities" (OuterVolumeSpecName: "utilities") pod "75d44e5f-cb22-4c92-8695-b9b8128d8724" (UID: "75d44e5f-cb22-4c92-8695-b9b8128d8724"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:49:31 crc kubenswrapper[4702]: I1125 13:49:31.676493 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75d44e5f-cb22-4c92-8695-b9b8128d8724-kube-api-access-4wncr" (OuterVolumeSpecName: "kube-api-access-4wncr") pod "75d44e5f-cb22-4c92-8695-b9b8128d8724" (UID: "75d44e5f-cb22-4c92-8695-b9b8128d8724"). InnerVolumeSpecName "kube-api-access-4wncr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:49:31 crc kubenswrapper[4702]: I1125 13:49:31.766507 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wncr\" (UniqueName: \"kubernetes.io/projected/75d44e5f-cb22-4c92-8695-b9b8128d8724-kube-api-access-4wncr\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:31 crc kubenswrapper[4702]: I1125 13:49:31.766551 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75d44e5f-cb22-4c92-8695-b9b8128d8724-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:31 crc kubenswrapper[4702]: I1125 13:49:31.810287 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3babd873-f61e-40d1-a42f-8a2f32f3ca40" path="/var/lib/kubelet/pods/3babd873-f61e-40d1-a42f-8a2f32f3ca40/volumes" Nov 25 13:49:31 crc kubenswrapper[4702]: I1125 13:49:31.811189 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a78d7c57-6155-429a-bbcf-168c3fa26e2e" path="/var/lib/kubelet/pods/a78d7c57-6155-429a-bbcf-168c3fa26e2e/volumes" Nov 25 13:49:32 crc kubenswrapper[4702]: I1125 13:49:32.191935 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5xb7k" Nov 25 13:49:32 crc kubenswrapper[4702]: I1125 13:49:32.440011 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rh58q" event={"ID":"75d44e5f-cb22-4c92-8695-b9b8128d8724","Type":"ContainerDied","Data":"4aebbab9169ca54a57e4034f63a82d7d371f38cf0ac5666265cc838c171620ce"} Nov 25 13:49:32 crc kubenswrapper[4702]: I1125 13:49:32.440089 4702 scope.go:117] "RemoveContainer" containerID="933261426a8ef97c9c22eac32e69ef98a98c29682be91f129fc80bbcce1ab3c2" Nov 25 13:49:32 crc kubenswrapper[4702]: I1125 13:49:32.440308 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rh58q" Nov 25 13:49:32 crc kubenswrapper[4702]: I1125 13:49:32.460449 4702 scope.go:117] "RemoveContainer" containerID="41fbbcc18094cd39841affd6f742d142bb78f167c84e731525d5f82fbfb9c7d5" Nov 25 13:49:32 crc kubenswrapper[4702]: I1125 13:49:32.477225 4702 scope.go:117] "RemoveContainer" containerID="2349a88076ca64f27e2459432bcc42ee0349d91fd7626e82f04950b44b9cbc8e" Nov 25 13:49:32 crc kubenswrapper[4702]: I1125 13:49:32.710190 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75d44e5f-cb22-4c92-8695-b9b8128d8724-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "75d44e5f-cb22-4c92-8695-b9b8128d8724" (UID: "75d44e5f-cb22-4c92-8695-b9b8128d8724"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:49:32 crc kubenswrapper[4702]: I1125 13:49:32.770349 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rh58q"] Nov 25 13:49:32 crc kubenswrapper[4702]: I1125 13:49:32.772838 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rh58q"] Nov 25 13:49:32 crc kubenswrapper[4702]: I1125 13:49:32.778554 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75d44e5f-cb22-4c92-8695-b9b8128d8724-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:33 crc kubenswrapper[4702]: I1125 13:49:33.812282 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75d44e5f-cb22-4c92-8695-b9b8128d8724" path="/var/lib/kubelet/pods/75d44e5f-cb22-4c92-8695-b9b8128d8724/volumes" Nov 25 13:49:33 crc kubenswrapper[4702]: I1125 13:49:33.962701 4702 patch_prober.go:28] interesting pod/machine-config-daemon-qqr76 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 13:49:33 crc kubenswrapper[4702]: I1125 13:49:33.962779 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 13:49:34 crc kubenswrapper[4702]: I1125 13:49:34.631563 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5xb7k"] Nov 25 13:49:34 crc kubenswrapper[4702]: I1125 13:49:34.631786 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5xb7k" podUID="c48d725b-8394-44f2-90e4-7ba6576fbd77" containerName="registry-server" containerID="cri-o://9928413bfabd1ba62692153a1e5bd900b6119e5be7254990ff92e93effa6113a" gracePeriod=2 Nov 25 13:49:34 crc kubenswrapper[4702]: I1125 13:49:34.997170 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5xb7k" Nov 25 13:49:35 crc kubenswrapper[4702]: I1125 13:49:35.107805 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6znlx\" (UniqueName: \"kubernetes.io/projected/c48d725b-8394-44f2-90e4-7ba6576fbd77-kube-api-access-6znlx\") pod \"c48d725b-8394-44f2-90e4-7ba6576fbd77\" (UID: \"c48d725b-8394-44f2-90e4-7ba6576fbd77\") " Nov 25 13:49:35 crc kubenswrapper[4702]: I1125 13:49:35.107990 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c48d725b-8394-44f2-90e4-7ba6576fbd77-utilities\") pod \"c48d725b-8394-44f2-90e4-7ba6576fbd77\" (UID: \"c48d725b-8394-44f2-90e4-7ba6576fbd77\") " Nov 25 13:49:35 crc kubenswrapper[4702]: I1125 13:49:35.108056 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c48d725b-8394-44f2-90e4-7ba6576fbd77-catalog-content\") pod \"c48d725b-8394-44f2-90e4-7ba6576fbd77\" (UID: \"c48d725b-8394-44f2-90e4-7ba6576fbd77\") " Nov 25 13:49:35 crc kubenswrapper[4702]: I1125 13:49:35.108661 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c48d725b-8394-44f2-90e4-7ba6576fbd77-utilities" (OuterVolumeSpecName: "utilities") pod "c48d725b-8394-44f2-90e4-7ba6576fbd77" (UID: "c48d725b-8394-44f2-90e4-7ba6576fbd77"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:49:35 crc kubenswrapper[4702]: I1125 13:49:35.108904 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c48d725b-8394-44f2-90e4-7ba6576fbd77-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:35 crc kubenswrapper[4702]: I1125 13:49:35.113392 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c48d725b-8394-44f2-90e4-7ba6576fbd77-kube-api-access-6znlx" (OuterVolumeSpecName: "kube-api-access-6znlx") pod "c48d725b-8394-44f2-90e4-7ba6576fbd77" (UID: "c48d725b-8394-44f2-90e4-7ba6576fbd77"). InnerVolumeSpecName "kube-api-access-6znlx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:49:35 crc kubenswrapper[4702]: I1125 13:49:35.155904 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c48d725b-8394-44f2-90e4-7ba6576fbd77-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c48d725b-8394-44f2-90e4-7ba6576fbd77" (UID: "c48d725b-8394-44f2-90e4-7ba6576fbd77"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:49:35 crc kubenswrapper[4702]: I1125 13:49:35.210777 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c48d725b-8394-44f2-90e4-7ba6576fbd77-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:35 crc kubenswrapper[4702]: I1125 13:49:35.210812 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6znlx\" (UniqueName: \"kubernetes.io/projected/c48d725b-8394-44f2-90e4-7ba6576fbd77-kube-api-access-6znlx\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:35 crc kubenswrapper[4702]: I1125 13:49:35.461935 4702 generic.go:334] "Generic (PLEG): container finished" podID="c48d725b-8394-44f2-90e4-7ba6576fbd77" containerID="9928413bfabd1ba62692153a1e5bd900b6119e5be7254990ff92e93effa6113a" exitCode=0 Nov 25 13:49:35 crc kubenswrapper[4702]: I1125 13:49:35.462015 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5xb7k" Nov 25 13:49:35 crc kubenswrapper[4702]: I1125 13:49:35.462014 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5xb7k" event={"ID":"c48d725b-8394-44f2-90e4-7ba6576fbd77","Type":"ContainerDied","Data":"9928413bfabd1ba62692153a1e5bd900b6119e5be7254990ff92e93effa6113a"} Nov 25 13:49:35 crc kubenswrapper[4702]: I1125 13:49:35.462492 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5xb7k" event={"ID":"c48d725b-8394-44f2-90e4-7ba6576fbd77","Type":"ContainerDied","Data":"2d9de4ac3da08660b0236131a82991bc738884e9d4d8764f178aedc43125e5cf"} Nov 25 13:49:35 crc kubenswrapper[4702]: I1125 13:49:35.462519 4702 scope.go:117] "RemoveContainer" containerID="9928413bfabd1ba62692153a1e5bd900b6119e5be7254990ff92e93effa6113a" Nov 25 13:49:35 crc kubenswrapper[4702]: I1125 13:49:35.482570 4702 scope.go:117] "RemoveContainer" containerID="bb703016b3f3e7ba0ed1fc7b5fbe384ad810e8504fab9d6ad9fe599070021619" Nov 25 13:49:35 crc kubenswrapper[4702]: I1125 13:49:35.506817 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5xb7k"] Nov 25 13:49:35 crc kubenswrapper[4702]: I1125 13:49:35.511249 4702 scope.go:117] "RemoveContainer" containerID="ef6eb6b37fb585832cecf57b30ced7e5104bb925b10b7e2f9cdf1ea63fee1c39" Nov 25 13:49:35 crc kubenswrapper[4702]: I1125 13:49:35.516089 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5xb7k"] Nov 25 13:49:35 crc kubenswrapper[4702]: I1125 13:49:35.526506 4702 scope.go:117] "RemoveContainer" containerID="9928413bfabd1ba62692153a1e5bd900b6119e5be7254990ff92e93effa6113a" Nov 25 13:49:35 crc kubenswrapper[4702]: E1125 13:49:35.526932 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9928413bfabd1ba62692153a1e5bd900b6119e5be7254990ff92e93effa6113a\": container with ID starting with 9928413bfabd1ba62692153a1e5bd900b6119e5be7254990ff92e93effa6113a not found: ID does not exist" containerID="9928413bfabd1ba62692153a1e5bd900b6119e5be7254990ff92e93effa6113a" Nov 25 13:49:35 crc kubenswrapper[4702]: I1125 13:49:35.526978 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9928413bfabd1ba62692153a1e5bd900b6119e5be7254990ff92e93effa6113a"} err="failed to get container status \"9928413bfabd1ba62692153a1e5bd900b6119e5be7254990ff92e93effa6113a\": rpc error: code = NotFound desc = could not find container \"9928413bfabd1ba62692153a1e5bd900b6119e5be7254990ff92e93effa6113a\": container with ID starting with 9928413bfabd1ba62692153a1e5bd900b6119e5be7254990ff92e93effa6113a not found: ID does not exist" Nov 25 13:49:35 crc kubenswrapper[4702]: I1125 13:49:35.527034 4702 scope.go:117] "RemoveContainer" containerID="bb703016b3f3e7ba0ed1fc7b5fbe384ad810e8504fab9d6ad9fe599070021619" Nov 25 13:49:35 crc kubenswrapper[4702]: E1125 13:49:35.527389 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb703016b3f3e7ba0ed1fc7b5fbe384ad810e8504fab9d6ad9fe599070021619\": container with ID starting with bb703016b3f3e7ba0ed1fc7b5fbe384ad810e8504fab9d6ad9fe599070021619 not found: ID does not exist" containerID="bb703016b3f3e7ba0ed1fc7b5fbe384ad810e8504fab9d6ad9fe599070021619" Nov 25 13:49:35 crc kubenswrapper[4702]: I1125 13:49:35.527419 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb703016b3f3e7ba0ed1fc7b5fbe384ad810e8504fab9d6ad9fe599070021619"} err="failed to get container status \"bb703016b3f3e7ba0ed1fc7b5fbe384ad810e8504fab9d6ad9fe599070021619\": rpc error: code = NotFound desc = could not find container \"bb703016b3f3e7ba0ed1fc7b5fbe384ad810e8504fab9d6ad9fe599070021619\": container with ID starting with bb703016b3f3e7ba0ed1fc7b5fbe384ad810e8504fab9d6ad9fe599070021619 not found: ID does not exist" Nov 25 13:49:35 crc kubenswrapper[4702]: I1125 13:49:35.527438 4702 scope.go:117] "RemoveContainer" containerID="ef6eb6b37fb585832cecf57b30ced7e5104bb925b10b7e2f9cdf1ea63fee1c39" Nov 25 13:49:35 crc kubenswrapper[4702]: E1125 13:49:35.527678 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef6eb6b37fb585832cecf57b30ced7e5104bb925b10b7e2f9cdf1ea63fee1c39\": container with ID starting with ef6eb6b37fb585832cecf57b30ced7e5104bb925b10b7e2f9cdf1ea63fee1c39 not found: ID does not exist" containerID="ef6eb6b37fb585832cecf57b30ced7e5104bb925b10b7e2f9cdf1ea63fee1c39" Nov 25 13:49:35 crc kubenswrapper[4702]: I1125 13:49:35.527704 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef6eb6b37fb585832cecf57b30ced7e5104bb925b10b7e2f9cdf1ea63fee1c39"} err="failed to get container status \"ef6eb6b37fb585832cecf57b30ced7e5104bb925b10b7e2f9cdf1ea63fee1c39\": rpc error: code = NotFound desc = could not find container \"ef6eb6b37fb585832cecf57b30ced7e5104bb925b10b7e2f9cdf1ea63fee1c39\": container with ID starting with ef6eb6b37fb585832cecf57b30ced7e5104bb925b10b7e2f9cdf1ea63fee1c39 not found: ID does not exist" Nov 25 13:49:35 crc kubenswrapper[4702]: I1125 13:49:35.808120 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c48d725b-8394-44f2-90e4-7ba6576fbd77" path="/var/lib/kubelet/pods/c48d725b-8394-44f2-90e4-7ba6576fbd77/volumes" Nov 25 13:49:38 crc kubenswrapper[4702]: I1125 13:49:38.833520 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 13:49:41 crc kubenswrapper[4702]: E1125 13:49:41.340337 4702 file.go:109] "Unable to process watch event" err="can't process config file \"/etc/kubernetes/manifests/kube-apiserver-startup-monitor-pod.yaml\": /etc/kubernetes/manifests/kube-apiserver-startup-monitor-pod.yaml: couldn't parse as pod(Object 'Kind' is missing in 'null'), please check config file" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.341840 4702 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 25 13:49:41 crc kubenswrapper[4702]: E1125 13:49:41.342110 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a78d7c57-6155-429a-bbcf-168c3fa26e2e" containerName="extract-utilities" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.342129 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="a78d7c57-6155-429a-bbcf-168c3fa26e2e" containerName="extract-utilities" Nov 25 13:49:41 crc kubenswrapper[4702]: E1125 13:49:41.342139 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a78d7c57-6155-429a-bbcf-168c3fa26e2e" containerName="registry-server" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.342145 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="a78d7c57-6155-429a-bbcf-168c3fa26e2e" containerName="registry-server" Nov 25 13:49:41 crc kubenswrapper[4702]: E1125 13:49:41.342155 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3babd873-f61e-40d1-a42f-8a2f32f3ca40" containerName="extract-utilities" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.342165 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="3babd873-f61e-40d1-a42f-8a2f32f3ca40" containerName="extract-utilities" Nov 25 13:49:41 crc kubenswrapper[4702]: E1125 13:49:41.342184 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c48d725b-8394-44f2-90e4-7ba6576fbd77" containerName="registry-server" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.342193 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="c48d725b-8394-44f2-90e4-7ba6576fbd77" containerName="registry-server" Nov 25 13:49:41 crc kubenswrapper[4702]: E1125 13:49:41.342205 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a78d7c57-6155-429a-bbcf-168c3fa26e2e" containerName="extract-content" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.342213 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="a78d7c57-6155-429a-bbcf-168c3fa26e2e" containerName="extract-content" Nov 25 13:49:41 crc kubenswrapper[4702]: E1125 13:49:41.342269 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75d44e5f-cb22-4c92-8695-b9b8128d8724" containerName="extract-content" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.342283 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="75d44e5f-cb22-4c92-8695-b9b8128d8724" containerName="extract-content" Nov 25 13:49:41 crc kubenswrapper[4702]: E1125 13:49:41.342294 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75d44e5f-cb22-4c92-8695-b9b8128d8724" containerName="registry-server" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.342349 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="75d44e5f-cb22-4c92-8695-b9b8128d8724" containerName="registry-server" Nov 25 13:49:41 crc kubenswrapper[4702]: E1125 13:49:41.342366 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3babd873-f61e-40d1-a42f-8a2f32f3ca40" containerName="registry-server" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.342375 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="3babd873-f61e-40d1-a42f-8a2f32f3ca40" containerName="registry-server" Nov 25 13:49:41 crc kubenswrapper[4702]: E1125 13:49:41.342388 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c48d725b-8394-44f2-90e4-7ba6576fbd77" containerName="extract-content" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.342396 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="c48d725b-8394-44f2-90e4-7ba6576fbd77" containerName="extract-content" Nov 25 13:49:41 crc kubenswrapper[4702]: E1125 13:49:41.342407 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75d44e5f-cb22-4c92-8695-b9b8128d8724" containerName="extract-utilities" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.342414 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="75d44e5f-cb22-4c92-8695-b9b8128d8724" containerName="extract-utilities" Nov 25 13:49:41 crc kubenswrapper[4702]: E1125 13:49:41.342423 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c48d725b-8394-44f2-90e4-7ba6576fbd77" containerName="extract-utilities" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.342431 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="c48d725b-8394-44f2-90e4-7ba6576fbd77" containerName="extract-utilities" Nov 25 13:49:41 crc kubenswrapper[4702]: E1125 13:49:41.342442 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3babd873-f61e-40d1-a42f-8a2f32f3ca40" containerName="extract-content" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.342449 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="3babd873-f61e-40d1-a42f-8a2f32f3ca40" containerName="extract-content" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.342566 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="a78d7c57-6155-429a-bbcf-168c3fa26e2e" containerName="registry-server" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.342586 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="75d44e5f-cb22-4c92-8695-b9b8128d8724" containerName="registry-server" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.342601 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="c48d725b-8394-44f2-90e4-7ba6576fbd77" containerName="registry-server" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.342613 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="3babd873-f61e-40d1-a42f-8a2f32f3ca40" containerName="registry-server" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.342989 4702 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.343223 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.343525 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec" gracePeriod=15 Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.343680 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0" gracePeriod=15 Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.343832 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef" gracePeriod=15 Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.343914 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b" gracePeriod=15 Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.343947 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057" gracePeriod=15 Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.347645 4702 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 13:49:41 crc kubenswrapper[4702]: E1125 13:49:41.347987 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.348007 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 25 13:49:41 crc kubenswrapper[4702]: E1125 13:49:41.348022 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.348031 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 25 13:49:41 crc kubenswrapper[4702]: E1125 13:49:41.348045 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.348053 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 25 13:49:41 crc kubenswrapper[4702]: E1125 13:49:41.348065 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.348073 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 13:49:41 crc kubenswrapper[4702]: E1125 13:49:41.348090 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.348100 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 25 13:49:41 crc kubenswrapper[4702]: E1125 13:49:41.348121 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.348131 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.348293 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.348312 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.348332 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.348350 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.348365 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 25 13:49:41 crc kubenswrapper[4702]: E1125 13:49:41.348503 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.348514 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.348650 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.383113 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.383390 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.383414 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.383434 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.383895 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.383912 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.383957 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.383987 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.391849 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.484826 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.484883 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.484907 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.484933 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.484969 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.484991 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.484998 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.485020 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.485024 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.485049 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.485083 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.485075 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.485179 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.485069 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.484966 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.485433 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.503141 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.505080 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.505832 4702 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef" exitCode=0 Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.505871 4702 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0" exitCode=0 Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.505902 4702 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b" exitCode=0 Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.505913 4702 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057" exitCode=2 Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.505955 4702 scope.go:117] "RemoveContainer" containerID="4ba1c22f2ccbe2295e54eb2ba275a0b9102b89b73c6eb8b207ec97fe9da477c4" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.581217 4702 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Readiness probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]log ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]api-openshift-apiserver-available ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]api-openshift-oauth-apiserver-available ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]informer-sync ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]poststarthook/start-apiserver-admission-initializer ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]poststarthook/openshift.io-api-request-count-filter ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]poststarthook/openshift.io-startkubeinformers ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]poststarthook/generic-apiserver-start-informers ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]poststarthook/priority-and-fairness-config-consumer ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]poststarthook/priority-and-fairness-filter ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]poststarthook/storage-object-count-tracker-hook ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]poststarthook/start-apiextensions-informers ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]poststarthook/start-apiextensions-controllers ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]poststarthook/crd-informer-synced ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]poststarthook/start-system-namespaces-controller ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]poststarthook/start-cluster-authentication-info-controller ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]poststarthook/start-legacy-token-tracking-controller ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]poststarthook/start-service-ip-repair-controllers ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]poststarthook/rbac/bootstrap-roles ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]poststarthook/scheduling/bootstrap-system-priority-classes ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]poststarthook/priority-and-fairness-config-producer ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]poststarthook/bootstrap-controller ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]poststarthook/start-kube-aggregator-informers ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]poststarthook/apiservice-status-local-available-controller ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]poststarthook/apiservice-status-remote-available-controller ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]poststarthook/apiservice-registration-controller ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]poststarthook/apiservice-wait-for-first-sync ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]poststarthook/apiservice-discovery-controller ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]poststarthook/kube-apiserver-autoregistration ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]autoregister-completion ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]poststarthook/apiservice-openapi-controller ok Nov 25 13:49:41 crc kubenswrapper[4702]: [+]poststarthook/apiservice-openapiv3-controller ok Nov 25 13:49:41 crc kubenswrapper[4702]: [-]shutdown failed: reason withheld Nov 25 13:49:41 crc kubenswrapper[4702]: readyz check failed Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.581357 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 13:49:41 crc kubenswrapper[4702]: I1125 13:49:41.690859 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 13:49:42 crc kubenswrapper[4702]: I1125 13:49:42.513573 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 13:49:42 crc kubenswrapper[4702]: I1125 13:49:42.516035 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"6ea4dc7200af29cebfd3e0b29cabcc9431eeceb8252a20eeb94361bfe3182605"} Nov 25 13:49:42 crc kubenswrapper[4702]: I1125 13:49:42.516080 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"a6efda7c1610899401b85a40b363b607d5b2aa744681efd11bfa499673e69d7a"} Nov 25 13:49:46 crc kubenswrapper[4702]: I1125 13:49:46.547547 4702 generic.go:334] "Generic (PLEG): container finished" podID="70ee4d8e-0a03-4949-8c82-dc84b9ae5699" containerID="39fec14b7d5465bc6e100123bdbeb78d28942068e22897c6aed71d85dc6af613" exitCode=0 Nov 25 13:49:46 crc kubenswrapper[4702]: I1125 13:49:46.547605 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"70ee4d8e-0a03-4949-8c82-dc84b9ae5699","Type":"ContainerDied","Data":"39fec14b7d5465bc6e100123bdbeb78d28942068e22897c6aed71d85dc6af613"} Nov 25 13:49:46 crc kubenswrapper[4702]: E1125 13:49:46.711822 4702 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.180:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187b441fa2a8cadb openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 13:49:41.709327067 +0000 UTC m=+190.871977517,LastTimestamp:2025-11-25 13:49:41.709327067 +0000 UTC m=+190.871977517,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 13:49:46 crc kubenswrapper[4702]: I1125 13:49:46.813447 4702 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:46 crc kubenswrapper[4702]: I1125 13:49:46.814229 4702 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:46 crc kubenswrapper[4702]: I1125 13:49:46.815218 4702 status_manager.go:851] "Failed to get status for pod" podUID="70ee4d8e-0a03-4949-8c82-dc84b9ae5699" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:46 crc kubenswrapper[4702]: I1125 13:49:46.815771 4702 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:47 crc kubenswrapper[4702]: I1125 13:49:47.828336 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 13:49:47 crc kubenswrapper[4702]: I1125 13:49:47.830195 4702 status_manager.go:851] "Failed to get status for pod" podUID="70ee4d8e-0a03-4949-8c82-dc84b9ae5699" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:47 crc kubenswrapper[4702]: I1125 13:49:47.831369 4702 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:47 crc kubenswrapper[4702]: I1125 13:49:47.973774 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/70ee4d8e-0a03-4949-8c82-dc84b9ae5699-var-lock\") pod \"70ee4d8e-0a03-4949-8c82-dc84b9ae5699\" (UID: \"70ee4d8e-0a03-4949-8c82-dc84b9ae5699\") " Nov 25 13:49:47 crc kubenswrapper[4702]: I1125 13:49:47.973845 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/70ee4d8e-0a03-4949-8c82-dc84b9ae5699-kube-api-access\") pod \"70ee4d8e-0a03-4949-8c82-dc84b9ae5699\" (UID: \"70ee4d8e-0a03-4949-8c82-dc84b9ae5699\") " Nov 25 13:49:47 crc kubenswrapper[4702]: I1125 13:49:47.973877 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/70ee4d8e-0a03-4949-8c82-dc84b9ae5699-kubelet-dir\") pod \"70ee4d8e-0a03-4949-8c82-dc84b9ae5699\" (UID: \"70ee4d8e-0a03-4949-8c82-dc84b9ae5699\") " Nov 25 13:49:47 crc kubenswrapper[4702]: I1125 13:49:47.973919 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/70ee4d8e-0a03-4949-8c82-dc84b9ae5699-var-lock" (OuterVolumeSpecName: "var-lock") pod "70ee4d8e-0a03-4949-8c82-dc84b9ae5699" (UID: "70ee4d8e-0a03-4949-8c82-dc84b9ae5699"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 13:49:47 crc kubenswrapper[4702]: I1125 13:49:47.974505 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/70ee4d8e-0a03-4949-8c82-dc84b9ae5699-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "70ee4d8e-0a03-4949-8c82-dc84b9ae5699" (UID: "70ee4d8e-0a03-4949-8c82-dc84b9ae5699"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 13:49:47 crc kubenswrapper[4702]: I1125 13:49:47.975705 4702 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/70ee4d8e-0a03-4949-8c82-dc84b9ae5699-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:47 crc kubenswrapper[4702]: I1125 13:49:47.975768 4702 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/70ee4d8e-0a03-4949-8c82-dc84b9ae5699-var-lock\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:47 crc kubenswrapper[4702]: I1125 13:49:47.981527 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70ee4d8e-0a03-4949-8c82-dc84b9ae5699-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "70ee4d8e-0a03-4949-8c82-dc84b9ae5699" (UID: "70ee4d8e-0a03-4949-8c82-dc84b9ae5699"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:49:48 crc kubenswrapper[4702]: I1125 13:49:48.076849 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/70ee4d8e-0a03-4949-8c82-dc84b9ae5699-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:48 crc kubenswrapper[4702]: I1125 13:49:48.565115 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"70ee4d8e-0a03-4949-8c82-dc84b9ae5699","Type":"ContainerDied","Data":"46dbcf4fc8f469bf842092cc4170f339734716afa08095f60cac4f78e9be70d5"} Nov 25 13:49:48 crc kubenswrapper[4702]: I1125 13:49:48.565714 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46dbcf4fc8f469bf842092cc4170f339734716afa08095f60cac4f78e9be70d5" Nov 25 13:49:48 crc kubenswrapper[4702]: I1125 13:49:48.565477 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 13:49:48 crc kubenswrapper[4702]: I1125 13:49:48.583071 4702 status_manager.go:851] "Failed to get status for pod" podUID="70ee4d8e-0a03-4949-8c82-dc84b9ae5699" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:48 crc kubenswrapper[4702]: I1125 13:49:48.583652 4702 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:48 crc kubenswrapper[4702]: I1125 13:49:48.735541 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 13:49:48 crc kubenswrapper[4702]: I1125 13:49:48.737331 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:49:48 crc kubenswrapper[4702]: I1125 13:49:48.738687 4702 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:48 crc kubenswrapper[4702]: I1125 13:49:48.739360 4702 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:48 crc kubenswrapper[4702]: I1125 13:49:48.739860 4702 status_manager.go:851] "Failed to get status for pod" podUID="70ee4d8e-0a03-4949-8c82-dc84b9ae5699" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:48 crc kubenswrapper[4702]: I1125 13:49:48.892402 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 25 13:49:48 crc kubenswrapper[4702]: I1125 13:49:48.892486 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 25 13:49:48 crc kubenswrapper[4702]: I1125 13:49:48.892521 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 25 13:49:48 crc kubenswrapper[4702]: I1125 13:49:48.892571 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 13:49:48 crc kubenswrapper[4702]: I1125 13:49:48.892648 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 13:49:48 crc kubenswrapper[4702]: I1125 13:49:48.892690 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 13:49:48 crc kubenswrapper[4702]: I1125 13:49:48.893288 4702 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:48 crc kubenswrapper[4702]: I1125 13:49:48.893311 4702 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:48 crc kubenswrapper[4702]: I1125 13:49:48.893325 4702 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:49 crc kubenswrapper[4702]: I1125 13:49:49.577128 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 13:49:49 crc kubenswrapper[4702]: I1125 13:49:49.578386 4702 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec" exitCode=0 Nov 25 13:49:49 crc kubenswrapper[4702]: I1125 13:49:49.578482 4702 scope.go:117] "RemoveContainer" containerID="b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef" Nov 25 13:49:49 crc kubenswrapper[4702]: I1125 13:49:49.578674 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:49:49 crc kubenswrapper[4702]: I1125 13:49:49.596485 4702 status_manager.go:851] "Failed to get status for pod" podUID="70ee4d8e-0a03-4949-8c82-dc84b9ae5699" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:49 crc kubenswrapper[4702]: I1125 13:49:49.597342 4702 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:49 crc kubenswrapper[4702]: I1125 13:49:49.597974 4702 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:49 crc kubenswrapper[4702]: I1125 13:49:49.602733 4702 scope.go:117] "RemoveContainer" containerID="38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0" Nov 25 13:49:49 crc kubenswrapper[4702]: I1125 13:49:49.620746 4702 scope.go:117] "RemoveContainer" containerID="740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b" Nov 25 13:49:49 crc kubenswrapper[4702]: I1125 13:49:49.635226 4702 scope.go:117] "RemoveContainer" containerID="2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057" Nov 25 13:49:49 crc kubenswrapper[4702]: I1125 13:49:49.651680 4702 scope.go:117] "RemoveContainer" containerID="752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec" Nov 25 13:49:49 crc kubenswrapper[4702]: I1125 13:49:49.671698 4702 scope.go:117] "RemoveContainer" containerID="71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92" Nov 25 13:49:49 crc kubenswrapper[4702]: I1125 13:49:49.695521 4702 scope.go:117] "RemoveContainer" containerID="b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef" Nov 25 13:49:49 crc kubenswrapper[4702]: E1125 13:49:49.696287 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\": container with ID starting with b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef not found: ID does not exist" containerID="b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef" Nov 25 13:49:49 crc kubenswrapper[4702]: I1125 13:49:49.696331 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef"} err="failed to get container status \"b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\": rpc error: code = NotFound desc = could not find container \"b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef\": container with ID starting with b8f4865e2c5bc91f6b2ea41713dd3446360f1edf85be19b5c8ab28a48d2c98ef not found: ID does not exist" Nov 25 13:49:49 crc kubenswrapper[4702]: I1125 13:49:49.696362 4702 scope.go:117] "RemoveContainer" containerID="38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0" Nov 25 13:49:49 crc kubenswrapper[4702]: E1125 13:49:49.696739 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\": container with ID starting with 38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0 not found: ID does not exist" containerID="38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0" Nov 25 13:49:49 crc kubenswrapper[4702]: I1125 13:49:49.696799 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0"} err="failed to get container status \"38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\": rpc error: code = NotFound desc = could not find container \"38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0\": container with ID starting with 38f1a4478eaf7ffe72868c946f4988ca07aa77f363bccd3d9c9c4885ebffdec0 not found: ID does not exist" Nov 25 13:49:49 crc kubenswrapper[4702]: I1125 13:49:49.696841 4702 scope.go:117] "RemoveContainer" containerID="740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b" Nov 25 13:49:49 crc kubenswrapper[4702]: E1125 13:49:49.697179 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\": container with ID starting with 740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b not found: ID does not exist" containerID="740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b" Nov 25 13:49:49 crc kubenswrapper[4702]: I1125 13:49:49.697211 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b"} err="failed to get container status \"740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\": rpc error: code = NotFound desc = could not find container \"740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b\": container with ID starting with 740363a0150c777f3134b16107397f23a659dffc786f2d104822be159608e46b not found: ID does not exist" Nov 25 13:49:49 crc kubenswrapper[4702]: I1125 13:49:49.697270 4702 scope.go:117] "RemoveContainer" containerID="2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057" Nov 25 13:49:49 crc kubenswrapper[4702]: E1125 13:49:49.697550 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\": container with ID starting with 2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057 not found: ID does not exist" containerID="2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057" Nov 25 13:49:49 crc kubenswrapper[4702]: I1125 13:49:49.697581 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057"} err="failed to get container status \"2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\": rpc error: code = NotFound desc = could not find container \"2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057\": container with ID starting with 2cd31ac4293241d024ee65633296fd85b7223c5a8a2509ea2519be1e79a30057 not found: ID does not exist" Nov 25 13:49:49 crc kubenswrapper[4702]: I1125 13:49:49.697604 4702 scope.go:117] "RemoveContainer" containerID="752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec" Nov 25 13:49:49 crc kubenswrapper[4702]: E1125 13:49:49.697971 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\": container with ID starting with 752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec not found: ID does not exist" containerID="752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec" Nov 25 13:49:49 crc kubenswrapper[4702]: I1125 13:49:49.698010 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec"} err="failed to get container status \"752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\": rpc error: code = NotFound desc = could not find container \"752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec\": container with ID starting with 752ac247a2f0de70fd5fa4b27c1434de9b046da9d9f89205f6400327f39b0aec not found: ID does not exist" Nov 25 13:49:49 crc kubenswrapper[4702]: I1125 13:49:49.698024 4702 scope.go:117] "RemoveContainer" containerID="71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92" Nov 25 13:49:49 crc kubenswrapper[4702]: E1125 13:49:49.698393 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\": container with ID starting with 71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92 not found: ID does not exist" containerID="71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92" Nov 25 13:49:49 crc kubenswrapper[4702]: I1125 13:49:49.698436 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92"} err="failed to get container status \"71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\": rpc error: code = NotFound desc = could not find container \"71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92\": container with ID starting with 71f090ed23a215fcb7f7fb22abfdb7c235f5f0ca84da6c9677b4ae0c6a6e6d92 not found: ID does not exist" Nov 25 13:49:49 crc kubenswrapper[4702]: I1125 13:49:49.811642 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Nov 25 13:49:50 crc kubenswrapper[4702]: E1125 13:49:50.448949 4702 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:50 crc kubenswrapper[4702]: E1125 13:49:50.449744 4702 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:50 crc kubenswrapper[4702]: E1125 13:49:50.450301 4702 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:50 crc kubenswrapper[4702]: E1125 13:49:50.450852 4702 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:50 crc kubenswrapper[4702]: E1125 13:49:50.451370 4702 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:50 crc kubenswrapper[4702]: I1125 13:49:50.451410 4702 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Nov 25 13:49:50 crc kubenswrapper[4702]: E1125 13:49:50.451635 4702 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="200ms" Nov 25 13:49:50 crc kubenswrapper[4702]: E1125 13:49:50.653688 4702 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="400ms" Nov 25 13:49:51 crc kubenswrapper[4702]: E1125 13:49:51.054927 4702 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="800ms" Nov 25 13:49:51 crc kubenswrapper[4702]: I1125 13:49:51.804662 4702 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:51 crc kubenswrapper[4702]: I1125 13:49:51.806577 4702 status_manager.go:851] "Failed to get status for pod" podUID="70ee4d8e-0a03-4949-8c82-dc84b9ae5699" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:51 crc kubenswrapper[4702]: E1125 13:49:51.856163 4702 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="1.6s" Nov 25 13:49:53 crc kubenswrapper[4702]: E1125 13:49:53.458056 4702 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="3.2s" Nov 25 13:49:54 crc kubenswrapper[4702]: I1125 13:49:54.931607 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" podUID="ab12e899-20e3-4755-a537-3e54c5761af6" containerName="oauth-openshift" containerID="cri-o://cad53363793354aee615f19f39ca2f7c28c1b98cc6c7bbffd094efc5dab27bd5" gracePeriod=15 Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.231035 4702 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-dfr9q container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.34:6443/healthz\": dial tcp 10.217.0.34:6443: connect: connection refused" start-of-body= Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.231164 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" podUID="ab12e899-20e3-4755-a537-3e54c5761af6" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.34:6443/healthz\": dial tcp 10.217.0.34:6443: connect: connection refused" Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.625696 4702 generic.go:334] "Generic (PLEG): container finished" podID="ab12e899-20e3-4755-a537-3e54c5761af6" containerID="cad53363793354aee615f19f39ca2f7c28c1b98cc6c7bbffd094efc5dab27bd5" exitCode=0 Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.625842 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" event={"ID":"ab12e899-20e3-4755-a537-3e54c5761af6","Type":"ContainerDied","Data":"cad53363793354aee615f19f39ca2f7c28c1b98cc6c7bbffd094efc5dab27bd5"} Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.802885 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.803546 4702 status_manager.go:851] "Failed to get status for pod" podUID="70ee4d8e-0a03-4949-8c82-dc84b9ae5699" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.804682 4702 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.811352 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.812209 4702 status_manager.go:851] "Failed to get status for pod" podUID="70ee4d8e-0a03-4949-8c82-dc84b9ae5699" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.812675 4702 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.812907 4702 status_manager.go:851] "Failed to get status for pod" podUID="ab12e899-20e3-4755-a537-3e54c5761af6" pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dfr9q\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.820708 4702 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="77def600-8d7d-496f-9f98-d67c0f30305a" Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.820750 4702 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="77def600-8d7d-496f-9f98-d67c0f30305a" Nov 25 13:49:55 crc kubenswrapper[4702]: E1125 13:49:55.821443 4702 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.822316 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.905654 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-service-ca\") pod \"ab12e899-20e3-4755-a537-3e54c5761af6\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.905715 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-session\") pod \"ab12e899-20e3-4755-a537-3e54c5761af6\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.905746 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-user-template-error\") pod \"ab12e899-20e3-4755-a537-3e54c5761af6\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.905763 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-trusted-ca-bundle\") pod \"ab12e899-20e3-4755-a537-3e54c5761af6\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.905779 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cd67b\" (UniqueName: \"kubernetes.io/projected/ab12e899-20e3-4755-a537-3e54c5761af6-kube-api-access-cd67b\") pod \"ab12e899-20e3-4755-a537-3e54c5761af6\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.905800 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-router-certs\") pod \"ab12e899-20e3-4755-a537-3e54c5761af6\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.905821 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-user-template-provider-selection\") pod \"ab12e899-20e3-4755-a537-3e54c5761af6\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.905844 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-user-template-login\") pod \"ab12e899-20e3-4755-a537-3e54c5761af6\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.905909 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-user-idp-0-file-data\") pod \"ab12e899-20e3-4755-a537-3e54c5761af6\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.905971 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-serving-cert\") pod \"ab12e899-20e3-4755-a537-3e54c5761af6\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.906003 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-cliconfig\") pod \"ab12e899-20e3-4755-a537-3e54c5761af6\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.906095 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ab12e899-20e3-4755-a537-3e54c5761af6-audit-policies\") pod \"ab12e899-20e3-4755-a537-3e54c5761af6\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.906123 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ab12e899-20e3-4755-a537-3e54c5761af6-audit-dir\") pod \"ab12e899-20e3-4755-a537-3e54c5761af6\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.906151 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-ocp-branding-template\") pod \"ab12e899-20e3-4755-a537-3e54c5761af6\" (UID: \"ab12e899-20e3-4755-a537-3e54c5761af6\") " Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.906936 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "ab12e899-20e3-4755-a537-3e54c5761af6" (UID: "ab12e899-20e3-4755-a537-3e54c5761af6"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.907014 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "ab12e899-20e3-4755-a537-3e54c5761af6" (UID: "ab12e899-20e3-4755-a537-3e54c5761af6"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.907107 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab12e899-20e3-4755-a537-3e54c5761af6-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "ab12e899-20e3-4755-a537-3e54c5761af6" (UID: "ab12e899-20e3-4755-a537-3e54c5761af6"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.907558 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab12e899-20e3-4755-a537-3e54c5761af6-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "ab12e899-20e3-4755-a537-3e54c5761af6" (UID: "ab12e899-20e3-4755-a537-3e54c5761af6"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.908150 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "ab12e899-20e3-4755-a537-3e54c5761af6" (UID: "ab12e899-20e3-4755-a537-3e54c5761af6"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.913241 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab12e899-20e3-4755-a537-3e54c5761af6-kube-api-access-cd67b" (OuterVolumeSpecName: "kube-api-access-cd67b") pod "ab12e899-20e3-4755-a537-3e54c5761af6" (UID: "ab12e899-20e3-4755-a537-3e54c5761af6"). InnerVolumeSpecName "kube-api-access-cd67b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.913735 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "ab12e899-20e3-4755-a537-3e54c5761af6" (UID: "ab12e899-20e3-4755-a537-3e54c5761af6"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.913991 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "ab12e899-20e3-4755-a537-3e54c5761af6" (UID: "ab12e899-20e3-4755-a537-3e54c5761af6"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.914089 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "ab12e899-20e3-4755-a537-3e54c5761af6" (UID: "ab12e899-20e3-4755-a537-3e54c5761af6"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.914283 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "ab12e899-20e3-4755-a537-3e54c5761af6" (UID: "ab12e899-20e3-4755-a537-3e54c5761af6"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.914427 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "ab12e899-20e3-4755-a537-3e54c5761af6" (UID: "ab12e899-20e3-4755-a537-3e54c5761af6"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.914623 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "ab12e899-20e3-4755-a537-3e54c5761af6" (UID: "ab12e899-20e3-4755-a537-3e54c5761af6"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.914901 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "ab12e899-20e3-4755-a537-3e54c5761af6" (UID: "ab12e899-20e3-4755-a537-3e54c5761af6"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:49:55 crc kubenswrapper[4702]: I1125 13:49:55.915338 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "ab12e899-20e3-4755-a537-3e54c5761af6" (UID: "ab12e899-20e3-4755-a537-3e54c5761af6"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:49:55 crc kubenswrapper[4702]: E1125 13:49:55.927997 4702 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.180:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187b441fa2a8cadb openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 13:49:41.709327067 +0000 UTC m=+190.871977517,LastTimestamp:2025-11-25 13:49:41.709327067 +0000 UTC m=+190.871977517,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.007525 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.007559 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.007571 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.007580 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.007591 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cd67b\" (UniqueName: \"kubernetes.io/projected/ab12e899-20e3-4755-a537-3e54c5761af6-kube-api-access-cd67b\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.007604 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.007614 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.007626 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.007638 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.007648 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.007657 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.007667 4702 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ab12e899-20e3-4755-a537-3e54c5761af6-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.007676 4702 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ab12e899-20e3-4755-a537-3e54c5761af6-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.007684 4702 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ab12e899-20e3-4755-a537-3e54c5761af6-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.637123 4702 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="b6649b8d33ec4cffb7c602db86bbbe0a60f0d02c39f112aa8ea7d89099b5c7c4" exitCode=0 Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.637203 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"b6649b8d33ec4cffb7c602db86bbbe0a60f0d02c39f112aa8ea7d89099b5c7c4"} Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.637254 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"2395ada05acd4b9b0ddac4a343c7e2d0cb66217d9b1f780fbea0a9cb7d44c2d8"} Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.637509 4702 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="77def600-8d7d-496f-9f98-d67c0f30305a" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.637523 4702 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="77def600-8d7d-496f-9f98-d67c0f30305a" Nov 25 13:49:56 crc kubenswrapper[4702]: E1125 13:49:56.638008 4702 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.638091 4702 status_manager.go:851] "Failed to get status for pod" podUID="ab12e899-20e3-4755-a537-3e54c5761af6" pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dfr9q\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.638581 4702 status_manager.go:851] "Failed to get status for pod" podUID="70ee4d8e-0a03-4949-8c82-dc84b9ae5699" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.639022 4702 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.639709 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" event={"ID":"ab12e899-20e3-4755-a537-3e54c5761af6","Type":"ContainerDied","Data":"19775841748b298f15c092f075ddca7b6a703cd7a5d1f4487b5c1408c0878b4c"} Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.639772 4702 scope.go:117] "RemoveContainer" containerID="cad53363793354aee615f19f39ca2f7c28c1b98cc6c7bbffd094efc5dab27bd5" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.639781 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.640947 4702 status_manager.go:851] "Failed to get status for pod" podUID="ab12e899-20e3-4755-a537-3e54c5761af6" pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dfr9q\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.641425 4702 status_manager.go:851] "Failed to get status for pod" podUID="70ee4d8e-0a03-4949-8c82-dc84b9ae5699" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.642660 4702 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.644555 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.644594 4702 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773" exitCode=1 Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.644617 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773"} Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.645095 4702 scope.go:117] "RemoveContainer" containerID="e7c7859ec3e2e80ccfdd874da9f42dfcd44bb41b6feeb8f3677a2dc1beba0773" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.646047 4702 status_manager.go:851] "Failed to get status for pod" podUID="ab12e899-20e3-4755-a537-3e54c5761af6" pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dfr9q\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.646335 4702 status_manager.go:851] "Failed to get status for pod" podUID="70ee4d8e-0a03-4949-8c82-dc84b9ae5699" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.646554 4702 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.647027 4702 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:56 crc kubenswrapper[4702]: E1125 13:49:56.663878 4702 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="6.4s" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.671769 4702 status_manager.go:851] "Failed to get status for pod" podUID="70ee4d8e-0a03-4949-8c82-dc84b9ae5699" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.672445 4702 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.673087 4702 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:56 crc kubenswrapper[4702]: I1125 13:49:56.673722 4702 status_manager.go:851] "Failed to get status for pod" podUID="ab12e899-20e3-4755-a537-3e54c5761af6" pod="openshift-authentication/oauth-openshift-558db77b4-dfr9q" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dfr9q\": dial tcp 38.102.83.180:6443: connect: connection refused" Nov 25 13:49:57 crc kubenswrapper[4702]: I1125 13:49:57.655429 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 25 13:49:57 crc kubenswrapper[4702]: I1125 13:49:57.655985 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2768877566a54299f965a9b48db907a634df961f20de03d6b220bda9c526b5cf"} Nov 25 13:49:57 crc kubenswrapper[4702]: I1125 13:49:57.658983 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"902e9cd2f169ea545a274678712e08fa09eb3ebc4f1a22446672a9c04ef20bd2"} Nov 25 13:49:58 crc kubenswrapper[4702]: I1125 13:49:58.670869 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"94c0bf2372d1ca74f809651f782a68645e8a7ddb7cd663f35ea44e869999cc12"} Nov 25 13:49:58 crc kubenswrapper[4702]: I1125 13:49:58.670915 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"95c7ce4439f52b07cd663af659d1372a430cf2d502cefb69d86a0e6eff6f345f"} Nov 25 13:49:58 crc kubenswrapper[4702]: I1125 13:49:58.670926 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d420334c0d67bf8761a6f652f85b5d9387b5cd7c3eccdcf424a83a7281ff18a0"} Nov 25 13:49:58 crc kubenswrapper[4702]: I1125 13:49:58.670937 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"7a0009e050aaaa0cb171d8570f3e9ecad5cf337164037b762b053b76f57d1548"} Nov 25 13:49:58 crc kubenswrapper[4702]: I1125 13:49:58.671112 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:49:58 crc kubenswrapper[4702]: I1125 13:49:58.671158 4702 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="77def600-8d7d-496f-9f98-d67c0f30305a" Nov 25 13:49:58 crc kubenswrapper[4702]: I1125 13:49:58.671182 4702 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="77def600-8d7d-496f-9f98-d67c0f30305a" Nov 25 13:50:00 crc kubenswrapper[4702]: I1125 13:50:00.308743 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 13:50:00 crc kubenswrapper[4702]: I1125 13:50:00.309019 4702 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Nov 25 13:50:00 crc kubenswrapper[4702]: I1125 13:50:00.309871 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Nov 25 13:50:00 crc kubenswrapper[4702]: I1125 13:50:00.823148 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:50:00 crc kubenswrapper[4702]: I1125 13:50:00.823292 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:50:00 crc kubenswrapper[4702]: I1125 13:50:00.833182 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:50:03 crc kubenswrapper[4702]: I1125 13:50:03.080483 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 13:50:03 crc kubenswrapper[4702]: I1125 13:50:03.964432 4702 patch_prober.go:28] interesting pod/machine-config-daemon-qqr76 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 13:50:03 crc kubenswrapper[4702]: I1125 13:50:03.964876 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 13:50:03 crc kubenswrapper[4702]: I1125 13:50:03.964943 4702 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" Nov 25 13:50:03 crc kubenswrapper[4702]: I1125 13:50:03.965928 4702 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a"} pod="openshift-machine-config-operator/machine-config-daemon-qqr76" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 13:50:03 crc kubenswrapper[4702]: I1125 13:50:03.966015 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" containerID="cri-o://1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a" gracePeriod=600 Nov 25 13:50:04 crc kubenswrapper[4702]: I1125 13:50:04.008631 4702 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:50:04 crc kubenswrapper[4702]: I1125 13:50:04.130304 4702 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="8aa260bd-458c-42a8-a51f-8988a6533a4e" Nov 25 13:50:04 crc kubenswrapper[4702]: I1125 13:50:04.720449 4702 generic.go:334] "Generic (PLEG): container finished" podID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerID="1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a" exitCode=0 Nov 25 13:50:04 crc kubenswrapper[4702]: I1125 13:50:04.721332 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" event={"ID":"3c491818-b0bb-4d82-a031-96e2dfac8c27","Type":"ContainerDied","Data":"1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a"} Nov 25 13:50:04 crc kubenswrapper[4702]: I1125 13:50:04.721412 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" event={"ID":"3c491818-b0bb-4d82-a031-96e2dfac8c27","Type":"ContainerStarted","Data":"9283d384600217f266003d80798e4dd2a8e101df13b5073b74f9e20b9c61f5a6"} Nov 25 13:50:04 crc kubenswrapper[4702]: I1125 13:50:04.721431 4702 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="77def600-8d7d-496f-9f98-d67c0f30305a" Nov 25 13:50:04 crc kubenswrapper[4702]: I1125 13:50:04.721452 4702 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="77def600-8d7d-496f-9f98-d67c0f30305a" Nov 25 13:50:04 crc kubenswrapper[4702]: I1125 13:50:04.731575 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:50:04 crc kubenswrapper[4702]: I1125 13:50:04.747980 4702 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="8aa260bd-458c-42a8-a51f-8988a6533a4e" Nov 25 13:50:05 crc kubenswrapper[4702]: I1125 13:50:05.726213 4702 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="77def600-8d7d-496f-9f98-d67c0f30305a" Nov 25 13:50:05 crc kubenswrapper[4702]: I1125 13:50:05.726645 4702 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="77def600-8d7d-496f-9f98-d67c0f30305a" Nov 25 13:50:05 crc kubenswrapper[4702]: I1125 13:50:05.730006 4702 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="8aa260bd-458c-42a8-a51f-8988a6533a4e" Nov 25 13:50:10 crc kubenswrapper[4702]: I1125 13:50:10.312193 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 13:50:10 crc kubenswrapper[4702]: I1125 13:50:10.317321 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 13:50:10 crc kubenswrapper[4702]: I1125 13:50:10.486313 4702 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 25 13:50:10 crc kubenswrapper[4702]: I1125 13:50:10.612941 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 25 13:50:10 crc kubenswrapper[4702]: I1125 13:50:10.713317 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 25 13:50:10 crc kubenswrapper[4702]: I1125 13:50:10.746154 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 13:50:10 crc kubenswrapper[4702]: I1125 13:50:10.806021 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 25 13:50:10 crc kubenswrapper[4702]: I1125 13:50:10.828830 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 25 13:50:11 crc kubenswrapper[4702]: I1125 13:50:11.240878 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 25 13:50:11 crc kubenswrapper[4702]: I1125 13:50:11.284065 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 25 13:50:11 crc kubenswrapper[4702]: I1125 13:50:11.319702 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 13:50:11 crc kubenswrapper[4702]: I1125 13:50:11.381121 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 25 13:50:11 crc kubenswrapper[4702]: I1125 13:50:11.418156 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 25 13:50:11 crc kubenswrapper[4702]: I1125 13:50:11.451103 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 25 13:50:11 crc kubenswrapper[4702]: I1125 13:50:11.591808 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 25 13:50:11 crc kubenswrapper[4702]: I1125 13:50:11.616837 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 25 13:50:11 crc kubenswrapper[4702]: I1125 13:50:11.669562 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 25 13:50:11 crc kubenswrapper[4702]: I1125 13:50:11.723536 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 25 13:50:11 crc kubenswrapper[4702]: I1125 13:50:11.822028 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 25 13:50:11 crc kubenswrapper[4702]: I1125 13:50:11.843734 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 25 13:50:12 crc kubenswrapper[4702]: I1125 13:50:12.128016 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 25 13:50:12 crc kubenswrapper[4702]: I1125 13:50:12.148145 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 25 13:50:12 crc kubenswrapper[4702]: I1125 13:50:12.337921 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 25 13:50:12 crc kubenswrapper[4702]: I1125 13:50:12.564578 4702 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 25 13:50:12 crc kubenswrapper[4702]: I1125 13:50:12.784141 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 25 13:50:12 crc kubenswrapper[4702]: I1125 13:50:12.816701 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 25 13:50:12 crc kubenswrapper[4702]: I1125 13:50:12.823643 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 25 13:50:12 crc kubenswrapper[4702]: I1125 13:50:12.870764 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 25 13:50:12 crc kubenswrapper[4702]: I1125 13:50:12.881985 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 13:50:12 crc kubenswrapper[4702]: I1125 13:50:12.884416 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 25 13:50:12 crc kubenswrapper[4702]: I1125 13:50:12.926540 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 25 13:50:13 crc kubenswrapper[4702]: I1125 13:50:13.034861 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 25 13:50:13 crc kubenswrapper[4702]: I1125 13:50:13.095561 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 25 13:50:13 crc kubenswrapper[4702]: I1125 13:50:13.099956 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 25 13:50:13 crc kubenswrapper[4702]: I1125 13:50:13.110574 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 25 13:50:13 crc kubenswrapper[4702]: I1125 13:50:13.199873 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 25 13:50:13 crc kubenswrapper[4702]: I1125 13:50:13.288284 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 25 13:50:13 crc kubenswrapper[4702]: I1125 13:50:13.301334 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 25 13:50:13 crc kubenswrapper[4702]: I1125 13:50:13.309901 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 25 13:50:13 crc kubenswrapper[4702]: I1125 13:50:13.489636 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 25 13:50:13 crc kubenswrapper[4702]: I1125 13:50:13.516947 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 25 13:50:13 crc kubenswrapper[4702]: I1125 13:50:13.653742 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 25 13:50:13 crc kubenswrapper[4702]: I1125 13:50:13.767710 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 25 13:50:13 crc kubenswrapper[4702]: I1125 13:50:13.826576 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 25 13:50:13 crc kubenswrapper[4702]: I1125 13:50:13.868349 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 25 13:50:13 crc kubenswrapper[4702]: I1125 13:50:13.889482 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 25 13:50:13 crc kubenswrapper[4702]: I1125 13:50:13.964795 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 25 13:50:13 crc kubenswrapper[4702]: I1125 13:50:13.985347 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 25 13:50:14 crc kubenswrapper[4702]: I1125 13:50:14.042954 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 25 13:50:14 crc kubenswrapper[4702]: I1125 13:50:14.181493 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 25 13:50:14 crc kubenswrapper[4702]: I1125 13:50:14.328825 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 25 13:50:14 crc kubenswrapper[4702]: I1125 13:50:14.497942 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 25 13:50:14 crc kubenswrapper[4702]: I1125 13:50:14.787224 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 25 13:50:14 crc kubenswrapper[4702]: I1125 13:50:14.817646 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 25 13:50:14 crc kubenswrapper[4702]: I1125 13:50:14.843548 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 25 13:50:14 crc kubenswrapper[4702]: I1125 13:50:14.995211 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 25 13:50:15 crc kubenswrapper[4702]: I1125 13:50:15.158391 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 25 13:50:15 crc kubenswrapper[4702]: I1125 13:50:15.160086 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 25 13:50:15 crc kubenswrapper[4702]: I1125 13:50:15.187436 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 25 13:50:15 crc kubenswrapper[4702]: I1125 13:50:15.235206 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 25 13:50:15 crc kubenswrapper[4702]: I1125 13:50:15.270542 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 25 13:50:15 crc kubenswrapper[4702]: I1125 13:50:15.394671 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 25 13:50:15 crc kubenswrapper[4702]: I1125 13:50:15.523957 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 13:50:15 crc kubenswrapper[4702]: I1125 13:50:15.589007 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 25 13:50:15 crc kubenswrapper[4702]: I1125 13:50:15.770158 4702 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 25 13:50:15 crc kubenswrapper[4702]: I1125 13:50:15.817876 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 25 13:50:15 crc kubenswrapper[4702]: I1125 13:50:15.936964 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 25 13:50:16 crc kubenswrapper[4702]: I1125 13:50:16.044705 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 25 13:50:16 crc kubenswrapper[4702]: I1125 13:50:16.170218 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 25 13:50:16 crc kubenswrapper[4702]: I1125 13:50:16.207656 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 25 13:50:16 crc kubenswrapper[4702]: I1125 13:50:16.291881 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 25 13:50:16 crc kubenswrapper[4702]: I1125 13:50:16.385678 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 25 13:50:16 crc kubenswrapper[4702]: I1125 13:50:16.536724 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 25 13:50:16 crc kubenswrapper[4702]: I1125 13:50:16.589747 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 25 13:50:16 crc kubenswrapper[4702]: I1125 13:50:16.608082 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 25 13:50:16 crc kubenswrapper[4702]: I1125 13:50:16.620334 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 25 13:50:16 crc kubenswrapper[4702]: I1125 13:50:16.669752 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 25 13:50:16 crc kubenswrapper[4702]: I1125 13:50:16.774369 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 25 13:50:16 crc kubenswrapper[4702]: I1125 13:50:16.801316 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 25 13:50:16 crc kubenswrapper[4702]: I1125 13:50:16.913650 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 25 13:50:16 crc kubenswrapper[4702]: I1125 13:50:16.936685 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 25 13:50:17 crc kubenswrapper[4702]: I1125 13:50:17.583037 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 25 13:50:17 crc kubenswrapper[4702]: I1125 13:50:17.631013 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 25 13:50:18 crc kubenswrapper[4702]: I1125 13:50:18.196944 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 25 13:50:18 crc kubenswrapper[4702]: I1125 13:50:18.279800 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 25 13:50:18 crc kubenswrapper[4702]: I1125 13:50:18.566843 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 25 13:50:18 crc kubenswrapper[4702]: I1125 13:50:18.579866 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 25 13:50:18 crc kubenswrapper[4702]: I1125 13:50:18.737596 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 25 13:50:19 crc kubenswrapper[4702]: I1125 13:50:19.074558 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 25 13:50:19 crc kubenswrapper[4702]: I1125 13:50:19.470805 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 25 13:50:19 crc kubenswrapper[4702]: I1125 13:50:19.826299 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 25 13:50:20 crc kubenswrapper[4702]: I1125 13:50:20.022734 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 25 13:50:20 crc kubenswrapper[4702]: I1125 13:50:20.103040 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 25 13:50:20 crc kubenswrapper[4702]: I1125 13:50:20.119701 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 25 13:50:20 crc kubenswrapper[4702]: I1125 13:50:20.138913 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 25 13:50:20 crc kubenswrapper[4702]: I1125 13:50:20.153373 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 13:50:20 crc kubenswrapper[4702]: I1125 13:50:20.294477 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 25 13:50:20 crc kubenswrapper[4702]: I1125 13:50:20.300107 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 25 13:50:20 crc kubenswrapper[4702]: I1125 13:50:20.335788 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 25 13:50:20 crc kubenswrapper[4702]: I1125 13:50:20.362292 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 25 13:50:20 crc kubenswrapper[4702]: I1125 13:50:20.859892 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 25 13:50:20 crc kubenswrapper[4702]: I1125 13:50:20.868799 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 25 13:50:20 crc kubenswrapper[4702]: I1125 13:50:20.920824 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 25 13:50:21 crc kubenswrapper[4702]: I1125 13:50:21.028301 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 25 13:50:21 crc kubenswrapper[4702]: I1125 13:50:21.039594 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 25 13:50:21 crc kubenswrapper[4702]: I1125 13:50:21.092716 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 25 13:50:21 crc kubenswrapper[4702]: I1125 13:50:21.164116 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 25 13:50:21 crc kubenswrapper[4702]: I1125 13:50:21.224634 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 25 13:50:21 crc kubenswrapper[4702]: I1125 13:50:21.383866 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 25 13:50:21 crc kubenswrapper[4702]: I1125 13:50:21.527155 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 25 13:50:21 crc kubenswrapper[4702]: I1125 13:50:21.567645 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 25 13:50:21 crc kubenswrapper[4702]: I1125 13:50:21.577774 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 25 13:50:21 crc kubenswrapper[4702]: I1125 13:50:21.776116 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 25 13:50:21 crc kubenswrapper[4702]: I1125 13:50:21.778144 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 25 13:50:21 crc kubenswrapper[4702]: I1125 13:50:21.797168 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 13:50:21 crc kubenswrapper[4702]: I1125 13:50:21.939796 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 25 13:50:21 crc kubenswrapper[4702]: I1125 13:50:21.967158 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 25 13:50:21 crc kubenswrapper[4702]: I1125 13:50:21.995508 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 13:50:22 crc kubenswrapper[4702]: I1125 13:50:22.037043 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 13:50:22 crc kubenswrapper[4702]: I1125 13:50:22.079121 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 25 13:50:22 crc kubenswrapper[4702]: I1125 13:50:22.086800 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 25 13:50:22 crc kubenswrapper[4702]: I1125 13:50:22.092169 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 25 13:50:22 crc kubenswrapper[4702]: I1125 13:50:22.230742 4702 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 25 13:50:22 crc kubenswrapper[4702]: I1125 13:50:22.272860 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 25 13:50:22 crc kubenswrapper[4702]: I1125 13:50:22.573958 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 25 13:50:23 crc kubenswrapper[4702]: I1125 13:50:23.209560 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 25 13:50:23 crc kubenswrapper[4702]: I1125 13:50:23.243600 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 25 13:50:23 crc kubenswrapper[4702]: I1125 13:50:23.284916 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 25 13:50:23 crc kubenswrapper[4702]: I1125 13:50:23.413906 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 25 13:50:23 crc kubenswrapper[4702]: I1125 13:50:23.424471 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 25 13:50:23 crc kubenswrapper[4702]: I1125 13:50:23.568522 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 25 13:50:23 crc kubenswrapper[4702]: I1125 13:50:23.626471 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 25 13:50:23 crc kubenswrapper[4702]: I1125 13:50:23.734671 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 25 13:50:23 crc kubenswrapper[4702]: I1125 13:50:23.778205 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 25 13:50:23 crc kubenswrapper[4702]: I1125 13:50:23.781484 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 25 13:50:23 crc kubenswrapper[4702]: I1125 13:50:23.831650 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 25 13:50:23 crc kubenswrapper[4702]: I1125 13:50:23.974698 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 25 13:50:24 crc kubenswrapper[4702]: I1125 13:50:24.018677 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 25 13:50:24 crc kubenswrapper[4702]: I1125 13:50:24.081524 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 25 13:50:24 crc kubenswrapper[4702]: I1125 13:50:24.375072 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 25 13:50:24 crc kubenswrapper[4702]: I1125 13:50:24.444195 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 25 13:50:24 crc kubenswrapper[4702]: I1125 13:50:24.541475 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.037882 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.176927 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.352380 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.412893 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.450328 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.591484 4702 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.595080 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=44.595048595 podStartE2EDuration="44.595048595s" podCreationTimestamp="2025-11-25 13:49:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:50:04.095194448 +0000 UTC m=+213.257844908" watchObservedRunningTime="2025-11-25 13:50:25.595048595 +0000 UTC m=+234.757699045" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.597749 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-dfr9q"] Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.597832 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-59978ccd59-lpl95"] Nov 25 13:50:25 crc kubenswrapper[4702]: E1125 13:50:25.598213 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab12e899-20e3-4755-a537-3e54c5761af6" containerName="oauth-openshift" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.598275 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab12e899-20e3-4755-a537-3e54c5761af6" containerName="oauth-openshift" Nov 25 13:50:25 crc kubenswrapper[4702]: E1125 13:50:25.598309 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70ee4d8e-0a03-4949-8c82-dc84b9ae5699" containerName="installer" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.598319 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="70ee4d8e-0a03-4949-8c82-dc84b9ae5699" containerName="installer" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.598390 4702 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="77def600-8d7d-496f-9f98-d67c0f30305a" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.598441 4702 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="77def600-8d7d-496f-9f98-d67c0f30305a" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.598463 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab12e899-20e3-4755-a537-3e54c5761af6" containerName="oauth-openshift" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.598481 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="70ee4d8e-0a03-4949-8c82-dc84b9ae5699" containerName="installer" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.599053 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.603780 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.605402 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.605732 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.605844 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.606088 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.606686 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.607115 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.607197 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.607226 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.607118 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.607328 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.607218 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.609521 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.620719 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.622769 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.634460 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.635830 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.655521 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=21.655500558 podStartE2EDuration="21.655500558s" podCreationTimestamp="2025-11-25 13:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:50:25.648023342 +0000 UTC m=+234.810673812" watchObservedRunningTime="2025-11-25 13:50:25.655500558 +0000 UTC m=+234.818151008" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.661806 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-audit-dir\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.661961 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.662122 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.662191 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxxdt\" (UniqueName: \"kubernetes.io/projected/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-kube-api-access-sxxdt\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.662253 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-audit-policies\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.662278 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.662330 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-v4-0-config-user-template-error\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.662356 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.662380 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-v4-0-config-system-service-ca\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.662409 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.662431 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-v4-0-config-user-template-login\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.662451 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-v4-0-config-system-session\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.662475 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.662501 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-v4-0-config-system-router-certs\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.729580 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.763589 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-audit-policies\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.764048 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.764163 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-v4-0-config-user-template-error\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.764306 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.764487 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-v4-0-config-system-service-ca\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.764582 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.764677 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-v4-0-config-user-template-login\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.764763 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-v4-0-config-system-session\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.764846 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.764931 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-v4-0-config-system-router-certs\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.765019 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-audit-dir\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.765114 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.765207 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.765316 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxxdt\" (UniqueName: \"kubernetes.io/projected/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-kube-api-access-sxxdt\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.765810 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.765810 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-v4-0-config-system-service-ca\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.764833 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-audit-policies\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.766079 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-audit-dir\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.766945 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.772687 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-v4-0-config-system-session\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.772802 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-v4-0-config-user-template-login\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.773521 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.774425 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-v4-0-config-user-template-error\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.775660 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.777403 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.778198 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.779143 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-v4-0-config-system-router-certs\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.793030 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxxdt\" (UniqueName: \"kubernetes.io/projected/77bda050-afab-48fe-a0e6-ddf6acf4ff0c-kube-api-access-sxxdt\") pod \"oauth-openshift-59978ccd59-lpl95\" (UID: \"77bda050-afab-48fe-a0e6-ddf6acf4ff0c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.810511 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab12e899-20e3-4755-a537-3e54c5761af6" path="/var/lib/kubelet/pods/ab12e899-20e3-4755-a537-3e54c5761af6/volumes" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.870888 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.924947 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:25 crc kubenswrapper[4702]: I1125 13:50:25.931606 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 25 13:50:26 crc kubenswrapper[4702]: I1125 13:50:26.019836 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 25 13:50:26 crc kubenswrapper[4702]: I1125 13:50:26.043947 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 25 13:50:26 crc kubenswrapper[4702]: I1125 13:50:26.065609 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 25 13:50:26 crc kubenswrapper[4702]: I1125 13:50:26.134392 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 25 13:50:26 crc kubenswrapper[4702]: I1125 13:50:26.212704 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 25 13:50:26 crc kubenswrapper[4702]: I1125 13:50:26.443079 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 25 13:50:26 crc kubenswrapper[4702]: I1125 13:50:26.497168 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 25 13:50:26 crc kubenswrapper[4702]: I1125 13:50:26.563035 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 25 13:50:26 crc kubenswrapper[4702]: I1125 13:50:26.572895 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 25 13:50:26 crc kubenswrapper[4702]: I1125 13:50:26.696313 4702 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 25 13:50:26 crc kubenswrapper[4702]: I1125 13:50:26.696611 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://6ea4dc7200af29cebfd3e0b29cabcc9431eeceb8252a20eeb94361bfe3182605" gracePeriod=5 Nov 25 13:50:26 crc kubenswrapper[4702]: I1125 13:50:26.703831 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 25 13:50:26 crc kubenswrapper[4702]: I1125 13:50:26.714656 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 25 13:50:26 crc kubenswrapper[4702]: I1125 13:50:26.765479 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 25 13:50:26 crc kubenswrapper[4702]: I1125 13:50:26.844302 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 25 13:50:26 crc kubenswrapper[4702]: I1125 13:50:26.873072 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 25 13:50:27 crc kubenswrapper[4702]: I1125 13:50:27.007893 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 25 13:50:27 crc kubenswrapper[4702]: I1125 13:50:27.043061 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 25 13:50:27 crc kubenswrapper[4702]: I1125 13:50:27.122693 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 25 13:50:27 crc kubenswrapper[4702]: I1125 13:50:27.203683 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 25 13:50:27 crc kubenswrapper[4702]: I1125 13:50:27.230953 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 25 13:50:27 crc kubenswrapper[4702]: I1125 13:50:27.243975 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 25 13:50:27 crc kubenswrapper[4702]: I1125 13:50:27.269527 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 25 13:50:27 crc kubenswrapper[4702]: I1125 13:50:27.304457 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 25 13:50:27 crc kubenswrapper[4702]: I1125 13:50:27.419424 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 13:50:27 crc kubenswrapper[4702]: I1125 13:50:27.441545 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 25 13:50:27 crc kubenswrapper[4702]: I1125 13:50:27.625784 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 25 13:50:27 crc kubenswrapper[4702]: I1125 13:50:27.626654 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 25 13:50:27 crc kubenswrapper[4702]: I1125 13:50:27.627250 4702 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 25 13:50:27 crc kubenswrapper[4702]: I1125 13:50:27.631606 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 25 13:50:27 crc kubenswrapper[4702]: I1125 13:50:27.786094 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 25 13:50:27 crc kubenswrapper[4702]: I1125 13:50:27.881915 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 25 13:50:27 crc kubenswrapper[4702]: I1125 13:50:27.914221 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 25 13:50:27 crc kubenswrapper[4702]: I1125 13:50:27.960523 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 25 13:50:28 crc kubenswrapper[4702]: I1125 13:50:28.107077 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 13:50:28 crc kubenswrapper[4702]: I1125 13:50:28.143560 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 25 13:50:28 crc kubenswrapper[4702]: I1125 13:50:28.180410 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 25 13:50:28 crc kubenswrapper[4702]: I1125 13:50:28.182569 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 13:50:28 crc kubenswrapper[4702]: I1125 13:50:28.186502 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 25 13:50:28 crc kubenswrapper[4702]: I1125 13:50:28.255356 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 25 13:50:28 crc kubenswrapper[4702]: I1125 13:50:28.370725 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 25 13:50:28 crc kubenswrapper[4702]: I1125 13:50:28.390077 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 25 13:50:28 crc kubenswrapper[4702]: I1125 13:50:28.557322 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 25 13:50:28 crc kubenswrapper[4702]: I1125 13:50:28.585546 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 25 13:50:28 crc kubenswrapper[4702]: I1125 13:50:28.599525 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 25 13:50:28 crc kubenswrapper[4702]: I1125 13:50:28.649559 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 25 13:50:28 crc kubenswrapper[4702]: I1125 13:50:28.769318 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 25 13:50:28 crc kubenswrapper[4702]: I1125 13:50:28.773943 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-59978ccd59-lpl95"] Nov 25 13:50:28 crc kubenswrapper[4702]: I1125 13:50:28.873178 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 25 13:50:28 crc kubenswrapper[4702]: I1125 13:50:28.881868 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 25 13:50:28 crc kubenswrapper[4702]: I1125 13:50:28.947461 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 25 13:50:28 crc kubenswrapper[4702]: I1125 13:50:28.955806 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 25 13:50:28 crc kubenswrapper[4702]: I1125 13:50:28.962558 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 25 13:50:28 crc kubenswrapper[4702]: I1125 13:50:28.979651 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-59978ccd59-lpl95"] Nov 25 13:50:29 crc kubenswrapper[4702]: I1125 13:50:29.075428 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 13:50:29 crc kubenswrapper[4702]: I1125 13:50:29.110993 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 25 13:50:29 crc kubenswrapper[4702]: I1125 13:50:29.163320 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 25 13:50:29 crc kubenswrapper[4702]: I1125 13:50:29.185203 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 25 13:50:29 crc kubenswrapper[4702]: I1125 13:50:29.197031 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 25 13:50:29 crc kubenswrapper[4702]: I1125 13:50:29.223534 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 25 13:50:29 crc kubenswrapper[4702]: I1125 13:50:29.265918 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 25 13:50:29 crc kubenswrapper[4702]: I1125 13:50:29.391718 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 25 13:50:29 crc kubenswrapper[4702]: I1125 13:50:29.432985 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 25 13:50:29 crc kubenswrapper[4702]: I1125 13:50:29.737272 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 25 13:50:29 crc kubenswrapper[4702]: I1125 13:50:29.777489 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 25 13:50:29 crc kubenswrapper[4702]: I1125 13:50:29.824536 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 25 13:50:29 crc kubenswrapper[4702]: I1125 13:50:29.915425 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" event={"ID":"77bda050-afab-48fe-a0e6-ddf6acf4ff0c","Type":"ContainerStarted","Data":"f761174c7e179b9b8e782cf05458ee2380d98131e810e8defe1ddf022ab4c263"} Nov 25 13:50:29 crc kubenswrapper[4702]: I1125 13:50:29.915743 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" event={"ID":"77bda050-afab-48fe-a0e6-ddf6acf4ff0c","Type":"ContainerStarted","Data":"ce7c4ccdac3c98de6e35b37904bb72237ac3eedbfc9ec1847226e382195ee54e"} Nov 25 13:50:29 crc kubenswrapper[4702]: I1125 13:50:29.916280 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:29 crc kubenswrapper[4702]: I1125 13:50:29.941761 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" Nov 25 13:50:29 crc kubenswrapper[4702]: I1125 13:50:29.956762 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-59978ccd59-lpl95" podStartSLOduration=60.956730999 podStartE2EDuration="1m0.956730999s" podCreationTimestamp="2025-11-25 13:49:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:50:29.955128146 +0000 UTC m=+239.117778596" watchObservedRunningTime="2025-11-25 13:50:29.956730999 +0000 UTC m=+239.119381449" Nov 25 13:50:30 crc kubenswrapper[4702]: I1125 13:50:30.058974 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 25 13:50:30 crc kubenswrapper[4702]: I1125 13:50:30.197346 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 25 13:50:30 crc kubenswrapper[4702]: I1125 13:50:30.255362 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 25 13:50:30 crc kubenswrapper[4702]: I1125 13:50:30.378683 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 25 13:50:30 crc kubenswrapper[4702]: I1125 13:50:30.533423 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 25 13:50:30 crc kubenswrapper[4702]: I1125 13:50:30.623548 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 25 13:50:30 crc kubenswrapper[4702]: I1125 13:50:30.726985 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 25 13:50:30 crc kubenswrapper[4702]: I1125 13:50:30.778313 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 25 13:50:30 crc kubenswrapper[4702]: I1125 13:50:30.814793 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 25 13:50:30 crc kubenswrapper[4702]: I1125 13:50:30.883154 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 25 13:50:31 crc kubenswrapper[4702]: I1125 13:50:31.025861 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 25 13:50:31 crc kubenswrapper[4702]: I1125 13:50:31.065354 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 25 13:50:31 crc kubenswrapper[4702]: I1125 13:50:31.065681 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 25 13:50:31 crc kubenswrapper[4702]: I1125 13:50:31.145825 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 25 13:50:31 crc kubenswrapper[4702]: I1125 13:50:31.876694 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 25 13:50:31 crc kubenswrapper[4702]: I1125 13:50:31.925899 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 25 13:50:31 crc kubenswrapper[4702]: I1125 13:50:31.926191 4702 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="6ea4dc7200af29cebfd3e0b29cabcc9431eeceb8252a20eeb94361bfe3182605" exitCode=137 Nov 25 13:50:32 crc kubenswrapper[4702]: I1125 13:50:32.057871 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 25 13:50:32 crc kubenswrapper[4702]: I1125 13:50:32.267168 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 25 13:50:32 crc kubenswrapper[4702]: I1125 13:50:32.294774 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 25 13:50:32 crc kubenswrapper[4702]: I1125 13:50:32.294839 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 13:50:32 crc kubenswrapper[4702]: I1125 13:50:32.458327 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 13:50:32 crc kubenswrapper[4702]: I1125 13:50:32.458429 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 13:50:32 crc kubenswrapper[4702]: I1125 13:50:32.458613 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 13:50:32 crc kubenswrapper[4702]: I1125 13:50:32.458691 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 13:50:32 crc kubenswrapper[4702]: I1125 13:50:32.458733 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 13:50:32 crc kubenswrapper[4702]: I1125 13:50:32.458803 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 13:50:32 crc kubenswrapper[4702]: I1125 13:50:32.458877 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 13:50:32 crc kubenswrapper[4702]: I1125 13:50:32.458909 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 13:50:32 crc kubenswrapper[4702]: I1125 13:50:32.459038 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 13:50:32 crc kubenswrapper[4702]: I1125 13:50:32.460341 4702 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:32 crc kubenswrapper[4702]: I1125 13:50:32.460362 4702 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:32 crc kubenswrapper[4702]: I1125 13:50:32.460371 4702 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:32 crc kubenswrapper[4702]: I1125 13:50:32.460380 4702 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:32 crc kubenswrapper[4702]: I1125 13:50:32.467016 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 13:50:32 crc kubenswrapper[4702]: I1125 13:50:32.487776 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 13:50:32 crc kubenswrapper[4702]: I1125 13:50:32.560866 4702 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:32 crc kubenswrapper[4702]: I1125 13:50:32.643953 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 25 13:50:32 crc kubenswrapper[4702]: I1125 13:50:32.751742 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 25 13:50:32 crc kubenswrapper[4702]: I1125 13:50:32.932415 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 25 13:50:32 crc kubenswrapper[4702]: I1125 13:50:32.932504 4702 scope.go:117] "RemoveContainer" containerID="6ea4dc7200af29cebfd3e0b29cabcc9431eeceb8252a20eeb94361bfe3182605" Nov 25 13:50:32 crc kubenswrapper[4702]: I1125 13:50:32.932671 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 13:50:32 crc kubenswrapper[4702]: I1125 13:50:32.970873 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 25 13:50:33 crc kubenswrapper[4702]: I1125 13:50:33.263602 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 25 13:50:33 crc kubenswrapper[4702]: I1125 13:50:33.810006 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Nov 25 13:50:33 crc kubenswrapper[4702]: I1125 13:50:33.810845 4702 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Nov 25 13:50:33 crc kubenswrapper[4702]: I1125 13:50:33.822727 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 25 13:50:33 crc kubenswrapper[4702]: I1125 13:50:33.822764 4702 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="a28d8346-e01a-4218-bd21-ce41ae392d04" Nov 25 13:50:33 crc kubenswrapper[4702]: I1125 13:50:33.826702 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 25 13:50:33 crc kubenswrapper[4702]: I1125 13:50:33.826736 4702 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="a28d8346-e01a-4218-bd21-ce41ae392d04" Nov 25 13:50:34 crc kubenswrapper[4702]: I1125 13:50:34.122116 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 13:50:34 crc kubenswrapper[4702]: I1125 13:50:34.151006 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 25 13:50:34 crc kubenswrapper[4702]: I1125 13:50:34.389446 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.300038 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pwlcj"] Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.301364 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pwlcj" podUID="66c1bf9b-3a89-427b-9bba-009bdafcfcc8" containerName="registry-server" containerID="cri-o://a8f0d3317038197b407ec5e59b13dc2e6109556f4b200e95589639b0b6630fd1" gracePeriod=30 Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.306595 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v9skf"] Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.306860 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-v9skf" podUID="597c2ecd-d536-4eec-bc02-89c798cbf25c" containerName="registry-server" containerID="cri-o://de2131333c95abcea4d33eff4ac4f0c9854933323f5b386caee89bf5fe159ee3" gracePeriod=30 Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.333466 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-zqrgn"] Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.333873 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-zqrgn" podUID="0389edfa-9a86-48df-a9a6-26e2b4d092c4" containerName="marketplace-operator" containerID="cri-o://1ddbe9a5a1c75085ea60f719d36e9a81327d12f38b34f6072fd550416ba6c8f8" gracePeriod=30 Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.337666 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jq6gh"] Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.338119 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jq6gh" podUID="f5b74e23-72c6-4278-9dc7-909f528d344b" containerName="registry-server" containerID="cri-o://9aeb281f96a9e4909ba74fcb0246b38a51b1578cf75be09905077750c19b1eb1" gracePeriod=30 Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.341256 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ppzc7"] Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.341616 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ppzc7" podUID="68f8eda2-8222-4ecb-a2d0-316ed8398d58" containerName="registry-server" containerID="cri-o://93047c8e45a295c8afe15bafe26622ba661947c901b3e30712219f6388c35216" gracePeriod=30 Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.350966 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8mh98"] Nov 25 13:50:37 crc kubenswrapper[4702]: E1125 13:50:37.351302 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.351320 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.351816 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.352329 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8mh98" Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.372402 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8mh98"] Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.528642 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b49d10e5-028c-4ad8-b03f-882744920b74-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8mh98\" (UID: \"b49d10e5-028c-4ad8-b03f-882744920b74\") " pod="openshift-marketplace/marketplace-operator-79b997595-8mh98" Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.528732 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b49d10e5-028c-4ad8-b03f-882744920b74-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8mh98\" (UID: \"b49d10e5-028c-4ad8-b03f-882744920b74\") " pod="openshift-marketplace/marketplace-operator-79b997595-8mh98" Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.528765 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh57s\" (UniqueName: \"kubernetes.io/projected/b49d10e5-028c-4ad8-b03f-882744920b74-kube-api-access-lh57s\") pod \"marketplace-operator-79b997595-8mh98\" (UID: \"b49d10e5-028c-4ad8-b03f-882744920b74\") " pod="openshift-marketplace/marketplace-operator-79b997595-8mh98" Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.630212 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b49d10e5-028c-4ad8-b03f-882744920b74-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8mh98\" (UID: \"b49d10e5-028c-4ad8-b03f-882744920b74\") " pod="openshift-marketplace/marketplace-operator-79b997595-8mh98" Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.630358 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b49d10e5-028c-4ad8-b03f-882744920b74-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8mh98\" (UID: \"b49d10e5-028c-4ad8-b03f-882744920b74\") " pod="openshift-marketplace/marketplace-operator-79b997595-8mh98" Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.630405 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh57s\" (UniqueName: \"kubernetes.io/projected/b49d10e5-028c-4ad8-b03f-882744920b74-kube-api-access-lh57s\") pod \"marketplace-operator-79b997595-8mh98\" (UID: \"b49d10e5-028c-4ad8-b03f-882744920b74\") " pod="openshift-marketplace/marketplace-operator-79b997595-8mh98" Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.633578 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b49d10e5-028c-4ad8-b03f-882744920b74-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8mh98\" (UID: \"b49d10e5-028c-4ad8-b03f-882744920b74\") " pod="openshift-marketplace/marketplace-operator-79b997595-8mh98" Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.640034 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b49d10e5-028c-4ad8-b03f-882744920b74-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8mh98\" (UID: \"b49d10e5-028c-4ad8-b03f-882744920b74\") " pod="openshift-marketplace/marketplace-operator-79b997595-8mh98" Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.661888 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh57s\" (UniqueName: \"kubernetes.io/projected/b49d10e5-028c-4ad8-b03f-882744920b74-kube-api-access-lh57s\") pod \"marketplace-operator-79b997595-8mh98\" (UID: \"b49d10e5-028c-4ad8-b03f-882744920b74\") " pod="openshift-marketplace/marketplace-operator-79b997595-8mh98" Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.826129 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8mh98" Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.840148 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-zqrgn" Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.849213 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ppzc7" Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.862022 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jq6gh" Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.892667 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pwlcj" Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.923419 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v9skf" Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.978883 4702 generic.go:334] "Generic (PLEG): container finished" podID="597c2ecd-d536-4eec-bc02-89c798cbf25c" containerID="de2131333c95abcea4d33eff4ac4f0c9854933323f5b386caee89bf5fe159ee3" exitCode=0 Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.978956 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v9skf" event={"ID":"597c2ecd-d536-4eec-bc02-89c798cbf25c","Type":"ContainerDied","Data":"de2131333c95abcea4d33eff4ac4f0c9854933323f5b386caee89bf5fe159ee3"} Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.978993 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v9skf" event={"ID":"597c2ecd-d536-4eec-bc02-89c798cbf25c","Type":"ContainerDied","Data":"77323b03247449a75a17a0a44b6b13d98acc7d342ad48da6c4b13244d733446c"} Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.979014 4702 scope.go:117] "RemoveContainer" containerID="de2131333c95abcea4d33eff4ac4f0c9854933323f5b386caee89bf5fe159ee3" Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.979190 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v9skf" Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.985583 4702 generic.go:334] "Generic (PLEG): container finished" podID="66c1bf9b-3a89-427b-9bba-009bdafcfcc8" containerID="a8f0d3317038197b407ec5e59b13dc2e6109556f4b200e95589639b0b6630fd1" exitCode=0 Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.985669 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pwlcj" event={"ID":"66c1bf9b-3a89-427b-9bba-009bdafcfcc8","Type":"ContainerDied","Data":"a8f0d3317038197b407ec5e59b13dc2e6109556f4b200e95589639b0b6630fd1"} Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.985709 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pwlcj" event={"ID":"66c1bf9b-3a89-427b-9bba-009bdafcfcc8","Type":"ContainerDied","Data":"582d5036f631fbdb494a4941759f3a9a42629120c463523af17627105cbe0d8e"} Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.985791 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pwlcj" Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.991826 4702 generic.go:334] "Generic (PLEG): container finished" podID="f5b74e23-72c6-4278-9dc7-909f528d344b" containerID="9aeb281f96a9e4909ba74fcb0246b38a51b1578cf75be09905077750c19b1eb1" exitCode=0 Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.991909 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jq6gh" event={"ID":"f5b74e23-72c6-4278-9dc7-909f528d344b","Type":"ContainerDied","Data":"9aeb281f96a9e4909ba74fcb0246b38a51b1578cf75be09905077750c19b1eb1"} Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.991939 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jq6gh" event={"ID":"f5b74e23-72c6-4278-9dc7-909f528d344b","Type":"ContainerDied","Data":"d6426bd7473101ce6ed7cf2ab5c1d3d453559de0441ee92b877072211873e9d4"} Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.991970 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jq6gh" Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.993487 4702 generic.go:334] "Generic (PLEG): container finished" podID="0389edfa-9a86-48df-a9a6-26e2b4d092c4" containerID="1ddbe9a5a1c75085ea60f719d36e9a81327d12f38b34f6072fd550416ba6c8f8" exitCode=0 Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.993551 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-zqrgn" event={"ID":"0389edfa-9a86-48df-a9a6-26e2b4d092c4","Type":"ContainerDied","Data":"1ddbe9a5a1c75085ea60f719d36e9a81327d12f38b34f6072fd550416ba6c8f8"} Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.993574 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-zqrgn" event={"ID":"0389edfa-9a86-48df-a9a6-26e2b4d092c4","Type":"ContainerDied","Data":"1b59567bdfc4086d232bd13a1c3b486b274c4e0516721c3a81d9c98ca22ad672"} Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.993652 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-zqrgn" Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.995912 4702 generic.go:334] "Generic (PLEG): container finished" podID="68f8eda2-8222-4ecb-a2d0-316ed8398d58" containerID="93047c8e45a295c8afe15bafe26622ba661947c901b3e30712219f6388c35216" exitCode=0 Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.995941 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ppzc7" event={"ID":"68f8eda2-8222-4ecb-a2d0-316ed8398d58","Type":"ContainerDied","Data":"93047c8e45a295c8afe15bafe26622ba661947c901b3e30712219f6388c35216"} Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.995958 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ppzc7" event={"ID":"68f8eda2-8222-4ecb-a2d0-316ed8398d58","Type":"ContainerDied","Data":"ed80ea5f96d6c387b4069090ffeb929bd2ca6dd1b70449cc88264b8a31c9fdb4"} Nov 25 13:50:37 crc kubenswrapper[4702]: I1125 13:50:37.996013 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ppzc7" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.001878 4702 scope.go:117] "RemoveContainer" containerID="899670a015b79035a701d6929e4b0dc3fb601ca8595069cc726ec84fc0befc9a" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.033675 4702 scope.go:117] "RemoveContainer" containerID="e7efc31fb65168ba89ae8b20e86503b7019eae41added19d7492b0e7d9058e5c" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.039312 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzwpq\" (UniqueName: \"kubernetes.io/projected/68f8eda2-8222-4ecb-a2d0-316ed8398d58-kube-api-access-jzwpq\") pod \"68f8eda2-8222-4ecb-a2d0-316ed8398d58\" (UID: \"68f8eda2-8222-4ecb-a2d0-316ed8398d58\") " Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.039382 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68f8eda2-8222-4ecb-a2d0-316ed8398d58-catalog-content\") pod \"68f8eda2-8222-4ecb-a2d0-316ed8398d58\" (UID: \"68f8eda2-8222-4ecb-a2d0-316ed8398d58\") " Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.039420 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5b4t\" (UniqueName: \"kubernetes.io/projected/66c1bf9b-3a89-427b-9bba-009bdafcfcc8-kube-api-access-l5b4t\") pod \"66c1bf9b-3a89-427b-9bba-009bdafcfcc8\" (UID: \"66c1bf9b-3a89-427b-9bba-009bdafcfcc8\") " Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.039450 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swzxn\" (UniqueName: \"kubernetes.io/projected/f5b74e23-72c6-4278-9dc7-909f528d344b-kube-api-access-swzxn\") pod \"f5b74e23-72c6-4278-9dc7-909f528d344b\" (UID: \"f5b74e23-72c6-4278-9dc7-909f528d344b\") " Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.039844 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66c1bf9b-3a89-427b-9bba-009bdafcfcc8-utilities\") pod \"66c1bf9b-3a89-427b-9bba-009bdafcfcc8\" (UID: \"66c1bf9b-3a89-427b-9bba-009bdafcfcc8\") " Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.039898 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/597c2ecd-d536-4eec-bc02-89c798cbf25c-catalog-content\") pod \"597c2ecd-d536-4eec-bc02-89c798cbf25c\" (UID: \"597c2ecd-d536-4eec-bc02-89c798cbf25c\") " Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.039934 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66c1bf9b-3a89-427b-9bba-009bdafcfcc8-catalog-content\") pod \"66c1bf9b-3a89-427b-9bba-009bdafcfcc8\" (UID: \"66c1bf9b-3a89-427b-9bba-009bdafcfcc8\") " Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.039972 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/597c2ecd-d536-4eec-bc02-89c798cbf25c-utilities\") pod \"597c2ecd-d536-4eec-bc02-89c798cbf25c\" (UID: \"597c2ecd-d536-4eec-bc02-89c798cbf25c\") " Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.040311 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5b74e23-72c6-4278-9dc7-909f528d344b-utilities\") pod \"f5b74e23-72c6-4278-9dc7-909f528d344b\" (UID: \"f5b74e23-72c6-4278-9dc7-909f528d344b\") " Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.040344 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0389edfa-9a86-48df-a9a6-26e2b4d092c4-marketplace-operator-metrics\") pod \"0389edfa-9a86-48df-a9a6-26e2b4d092c4\" (UID: \"0389edfa-9a86-48df-a9a6-26e2b4d092c4\") " Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.040373 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0389edfa-9a86-48df-a9a6-26e2b4d092c4-marketplace-trusted-ca\") pod \"0389edfa-9a86-48df-a9a6-26e2b4d092c4\" (UID: \"0389edfa-9a86-48df-a9a6-26e2b4d092c4\") " Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.040400 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ntsgr\" (UniqueName: \"kubernetes.io/projected/597c2ecd-d536-4eec-bc02-89c798cbf25c-kube-api-access-ntsgr\") pod \"597c2ecd-d536-4eec-bc02-89c798cbf25c\" (UID: \"597c2ecd-d536-4eec-bc02-89c798cbf25c\") " Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.040423 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vgk7\" (UniqueName: \"kubernetes.io/projected/0389edfa-9a86-48df-a9a6-26e2b4d092c4-kube-api-access-2vgk7\") pod \"0389edfa-9a86-48df-a9a6-26e2b4d092c4\" (UID: \"0389edfa-9a86-48df-a9a6-26e2b4d092c4\") " Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.040451 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5b74e23-72c6-4278-9dc7-909f528d344b-catalog-content\") pod \"f5b74e23-72c6-4278-9dc7-909f528d344b\" (UID: \"f5b74e23-72c6-4278-9dc7-909f528d344b\") " Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.040476 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68f8eda2-8222-4ecb-a2d0-316ed8398d58-utilities\") pod \"68f8eda2-8222-4ecb-a2d0-316ed8398d58\" (UID: \"68f8eda2-8222-4ecb-a2d0-316ed8398d58\") " Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.041133 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66c1bf9b-3a89-427b-9bba-009bdafcfcc8-utilities" (OuterVolumeSpecName: "utilities") pod "66c1bf9b-3a89-427b-9bba-009bdafcfcc8" (UID: "66c1bf9b-3a89-427b-9bba-009bdafcfcc8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.041390 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/597c2ecd-d536-4eec-bc02-89c798cbf25c-utilities" (OuterVolumeSpecName: "utilities") pod "597c2ecd-d536-4eec-bc02-89c798cbf25c" (UID: "597c2ecd-d536-4eec-bc02-89c798cbf25c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.041768 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68f8eda2-8222-4ecb-a2d0-316ed8398d58-utilities" (OuterVolumeSpecName: "utilities") pod "68f8eda2-8222-4ecb-a2d0-316ed8398d58" (UID: "68f8eda2-8222-4ecb-a2d0-316ed8398d58"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.043036 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5b74e23-72c6-4278-9dc7-909f528d344b-utilities" (OuterVolumeSpecName: "utilities") pod "f5b74e23-72c6-4278-9dc7-909f528d344b" (UID: "f5b74e23-72c6-4278-9dc7-909f528d344b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.043950 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0389edfa-9a86-48df-a9a6-26e2b4d092c4-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "0389edfa-9a86-48df-a9a6-26e2b4d092c4" (UID: "0389edfa-9a86-48df-a9a6-26e2b4d092c4"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.046193 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0389edfa-9a86-48df-a9a6-26e2b4d092c4-kube-api-access-2vgk7" (OuterVolumeSpecName: "kube-api-access-2vgk7") pod "0389edfa-9a86-48df-a9a6-26e2b4d092c4" (UID: "0389edfa-9a86-48df-a9a6-26e2b4d092c4"). InnerVolumeSpecName "kube-api-access-2vgk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.052117 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5b74e23-72c6-4278-9dc7-909f528d344b-kube-api-access-swzxn" (OuterVolumeSpecName: "kube-api-access-swzxn") pod "f5b74e23-72c6-4278-9dc7-909f528d344b" (UID: "f5b74e23-72c6-4278-9dc7-909f528d344b"). InnerVolumeSpecName "kube-api-access-swzxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.052146 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68f8eda2-8222-4ecb-a2d0-316ed8398d58-kube-api-access-jzwpq" (OuterVolumeSpecName: "kube-api-access-jzwpq") pod "68f8eda2-8222-4ecb-a2d0-316ed8398d58" (UID: "68f8eda2-8222-4ecb-a2d0-316ed8398d58"). InnerVolumeSpecName "kube-api-access-jzwpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.052518 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/597c2ecd-d536-4eec-bc02-89c798cbf25c-kube-api-access-ntsgr" (OuterVolumeSpecName: "kube-api-access-ntsgr") pod "597c2ecd-d536-4eec-bc02-89c798cbf25c" (UID: "597c2ecd-d536-4eec-bc02-89c798cbf25c"). InnerVolumeSpecName "kube-api-access-ntsgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.055888 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0389edfa-9a86-48df-a9a6-26e2b4d092c4-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "0389edfa-9a86-48df-a9a6-26e2b4d092c4" (UID: "0389edfa-9a86-48df-a9a6-26e2b4d092c4"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.057660 4702 scope.go:117] "RemoveContainer" containerID="de2131333c95abcea4d33eff4ac4f0c9854933323f5b386caee89bf5fe159ee3" Nov 25 13:50:38 crc kubenswrapper[4702]: E1125 13:50:38.059915 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de2131333c95abcea4d33eff4ac4f0c9854933323f5b386caee89bf5fe159ee3\": container with ID starting with de2131333c95abcea4d33eff4ac4f0c9854933323f5b386caee89bf5fe159ee3 not found: ID does not exist" containerID="de2131333c95abcea4d33eff4ac4f0c9854933323f5b386caee89bf5fe159ee3" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.059956 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de2131333c95abcea4d33eff4ac4f0c9854933323f5b386caee89bf5fe159ee3"} err="failed to get container status \"de2131333c95abcea4d33eff4ac4f0c9854933323f5b386caee89bf5fe159ee3\": rpc error: code = NotFound desc = could not find container \"de2131333c95abcea4d33eff4ac4f0c9854933323f5b386caee89bf5fe159ee3\": container with ID starting with de2131333c95abcea4d33eff4ac4f0c9854933323f5b386caee89bf5fe159ee3 not found: ID does not exist" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.059989 4702 scope.go:117] "RemoveContainer" containerID="899670a015b79035a701d6929e4b0dc3fb601ca8595069cc726ec84fc0befc9a" Nov 25 13:50:38 crc kubenswrapper[4702]: E1125 13:50:38.060468 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"899670a015b79035a701d6929e4b0dc3fb601ca8595069cc726ec84fc0befc9a\": container with ID starting with 899670a015b79035a701d6929e4b0dc3fb601ca8595069cc726ec84fc0befc9a not found: ID does not exist" containerID="899670a015b79035a701d6929e4b0dc3fb601ca8595069cc726ec84fc0befc9a" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.060532 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"899670a015b79035a701d6929e4b0dc3fb601ca8595069cc726ec84fc0befc9a"} err="failed to get container status \"899670a015b79035a701d6929e4b0dc3fb601ca8595069cc726ec84fc0befc9a\": rpc error: code = NotFound desc = could not find container \"899670a015b79035a701d6929e4b0dc3fb601ca8595069cc726ec84fc0befc9a\": container with ID starting with 899670a015b79035a701d6929e4b0dc3fb601ca8595069cc726ec84fc0befc9a not found: ID does not exist" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.060577 4702 scope.go:117] "RemoveContainer" containerID="e7efc31fb65168ba89ae8b20e86503b7019eae41added19d7492b0e7d9058e5c" Nov 25 13:50:38 crc kubenswrapper[4702]: E1125 13:50:38.060980 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7efc31fb65168ba89ae8b20e86503b7019eae41added19d7492b0e7d9058e5c\": container with ID starting with e7efc31fb65168ba89ae8b20e86503b7019eae41added19d7492b0e7d9058e5c not found: ID does not exist" containerID="e7efc31fb65168ba89ae8b20e86503b7019eae41added19d7492b0e7d9058e5c" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.061021 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7efc31fb65168ba89ae8b20e86503b7019eae41added19d7492b0e7d9058e5c"} err="failed to get container status \"e7efc31fb65168ba89ae8b20e86503b7019eae41added19d7492b0e7d9058e5c\": rpc error: code = NotFound desc = could not find container \"e7efc31fb65168ba89ae8b20e86503b7019eae41added19d7492b0e7d9058e5c\": container with ID starting with e7efc31fb65168ba89ae8b20e86503b7019eae41added19d7492b0e7d9058e5c not found: ID does not exist" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.061046 4702 scope.go:117] "RemoveContainer" containerID="a8f0d3317038197b407ec5e59b13dc2e6109556f4b200e95589639b0b6630fd1" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.064940 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66c1bf9b-3a89-427b-9bba-009bdafcfcc8-kube-api-access-l5b4t" (OuterVolumeSpecName: "kube-api-access-l5b4t") pod "66c1bf9b-3a89-427b-9bba-009bdafcfcc8" (UID: "66c1bf9b-3a89-427b-9bba-009bdafcfcc8"). InnerVolumeSpecName "kube-api-access-l5b4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.073724 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5b74e23-72c6-4278-9dc7-909f528d344b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f5b74e23-72c6-4278-9dc7-909f528d344b" (UID: "f5b74e23-72c6-4278-9dc7-909f528d344b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.081283 4702 scope.go:117] "RemoveContainer" containerID="324b02bfbcc635df3fc842caec60c5d6a61be5e70b0eeecff1af40baa8a1662f" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.089852 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8mh98"] Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.101585 4702 scope.go:117] "RemoveContainer" containerID="d18368f3544c7485a747ab6635a3dabf62583e82e595fd3004b7da8648b97179" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.114318 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66c1bf9b-3a89-427b-9bba-009bdafcfcc8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "66c1bf9b-3a89-427b-9bba-009bdafcfcc8" (UID: "66c1bf9b-3a89-427b-9bba-009bdafcfcc8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.122205 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/597c2ecd-d536-4eec-bc02-89c798cbf25c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "597c2ecd-d536-4eec-bc02-89c798cbf25c" (UID: "597c2ecd-d536-4eec-bc02-89c798cbf25c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.127869 4702 scope.go:117] "RemoveContainer" containerID="a8f0d3317038197b407ec5e59b13dc2e6109556f4b200e95589639b0b6630fd1" Nov 25 13:50:38 crc kubenswrapper[4702]: E1125 13:50:38.128522 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8f0d3317038197b407ec5e59b13dc2e6109556f4b200e95589639b0b6630fd1\": container with ID starting with a8f0d3317038197b407ec5e59b13dc2e6109556f4b200e95589639b0b6630fd1 not found: ID does not exist" containerID="a8f0d3317038197b407ec5e59b13dc2e6109556f4b200e95589639b0b6630fd1" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.128590 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8f0d3317038197b407ec5e59b13dc2e6109556f4b200e95589639b0b6630fd1"} err="failed to get container status \"a8f0d3317038197b407ec5e59b13dc2e6109556f4b200e95589639b0b6630fd1\": rpc error: code = NotFound desc = could not find container \"a8f0d3317038197b407ec5e59b13dc2e6109556f4b200e95589639b0b6630fd1\": container with ID starting with a8f0d3317038197b407ec5e59b13dc2e6109556f4b200e95589639b0b6630fd1 not found: ID does not exist" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.128636 4702 scope.go:117] "RemoveContainer" containerID="324b02bfbcc635df3fc842caec60c5d6a61be5e70b0eeecff1af40baa8a1662f" Nov 25 13:50:38 crc kubenswrapper[4702]: E1125 13:50:38.129276 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"324b02bfbcc635df3fc842caec60c5d6a61be5e70b0eeecff1af40baa8a1662f\": container with ID starting with 324b02bfbcc635df3fc842caec60c5d6a61be5e70b0eeecff1af40baa8a1662f not found: ID does not exist" containerID="324b02bfbcc635df3fc842caec60c5d6a61be5e70b0eeecff1af40baa8a1662f" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.129322 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"324b02bfbcc635df3fc842caec60c5d6a61be5e70b0eeecff1af40baa8a1662f"} err="failed to get container status \"324b02bfbcc635df3fc842caec60c5d6a61be5e70b0eeecff1af40baa8a1662f\": rpc error: code = NotFound desc = could not find container \"324b02bfbcc635df3fc842caec60c5d6a61be5e70b0eeecff1af40baa8a1662f\": container with ID starting with 324b02bfbcc635df3fc842caec60c5d6a61be5e70b0eeecff1af40baa8a1662f not found: ID does not exist" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.129358 4702 scope.go:117] "RemoveContainer" containerID="d18368f3544c7485a747ab6635a3dabf62583e82e595fd3004b7da8648b97179" Nov 25 13:50:38 crc kubenswrapper[4702]: E1125 13:50:38.129635 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d18368f3544c7485a747ab6635a3dabf62583e82e595fd3004b7da8648b97179\": container with ID starting with d18368f3544c7485a747ab6635a3dabf62583e82e595fd3004b7da8648b97179 not found: ID does not exist" containerID="d18368f3544c7485a747ab6635a3dabf62583e82e595fd3004b7da8648b97179" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.129669 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d18368f3544c7485a747ab6635a3dabf62583e82e595fd3004b7da8648b97179"} err="failed to get container status \"d18368f3544c7485a747ab6635a3dabf62583e82e595fd3004b7da8648b97179\": rpc error: code = NotFound desc = could not find container \"d18368f3544c7485a747ab6635a3dabf62583e82e595fd3004b7da8648b97179\": container with ID starting with d18368f3544c7485a747ab6635a3dabf62583e82e595fd3004b7da8648b97179 not found: ID does not exist" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.129687 4702 scope.go:117] "RemoveContainer" containerID="9aeb281f96a9e4909ba74fcb0246b38a51b1578cf75be09905077750c19b1eb1" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.141811 4702 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0389edfa-9a86-48df-a9a6-26e2b4d092c4-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.141848 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5b74e23-72c6-4278-9dc7-909f528d344b-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.141865 4702 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0389edfa-9a86-48df-a9a6-26e2b4d092c4-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.141879 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ntsgr\" (UniqueName: \"kubernetes.io/projected/597c2ecd-d536-4eec-bc02-89c798cbf25c-kube-api-access-ntsgr\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.141892 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vgk7\" (UniqueName: \"kubernetes.io/projected/0389edfa-9a86-48df-a9a6-26e2b4d092c4-kube-api-access-2vgk7\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.141906 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5b74e23-72c6-4278-9dc7-909f528d344b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.141919 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68f8eda2-8222-4ecb-a2d0-316ed8398d58-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.141932 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzwpq\" (UniqueName: \"kubernetes.io/projected/68f8eda2-8222-4ecb-a2d0-316ed8398d58-kube-api-access-jzwpq\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.141945 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5b4t\" (UniqueName: \"kubernetes.io/projected/66c1bf9b-3a89-427b-9bba-009bdafcfcc8-kube-api-access-l5b4t\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.141955 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swzxn\" (UniqueName: \"kubernetes.io/projected/f5b74e23-72c6-4278-9dc7-909f528d344b-kube-api-access-swzxn\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.141963 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66c1bf9b-3a89-427b-9bba-009bdafcfcc8-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.141973 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/597c2ecd-d536-4eec-bc02-89c798cbf25c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.141981 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66c1bf9b-3a89-427b-9bba-009bdafcfcc8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.141990 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/597c2ecd-d536-4eec-bc02-89c798cbf25c-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.145896 4702 scope.go:117] "RemoveContainer" containerID="92fdee327707bb0c612cfd0011379e14b1ff8ee7af7c4eebc8db3c471a4eb0b9" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.154994 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68f8eda2-8222-4ecb-a2d0-316ed8398d58-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "68f8eda2-8222-4ecb-a2d0-316ed8398d58" (UID: "68f8eda2-8222-4ecb-a2d0-316ed8398d58"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.167168 4702 scope.go:117] "RemoveContainer" containerID="052bbacfe3923529d43f8faf0dcca91bd7695404ebd9b87324946c534e265603" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.186896 4702 scope.go:117] "RemoveContainer" containerID="9aeb281f96a9e4909ba74fcb0246b38a51b1578cf75be09905077750c19b1eb1" Nov 25 13:50:38 crc kubenswrapper[4702]: E1125 13:50:38.187586 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9aeb281f96a9e4909ba74fcb0246b38a51b1578cf75be09905077750c19b1eb1\": container with ID starting with 9aeb281f96a9e4909ba74fcb0246b38a51b1578cf75be09905077750c19b1eb1 not found: ID does not exist" containerID="9aeb281f96a9e4909ba74fcb0246b38a51b1578cf75be09905077750c19b1eb1" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.187637 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9aeb281f96a9e4909ba74fcb0246b38a51b1578cf75be09905077750c19b1eb1"} err="failed to get container status \"9aeb281f96a9e4909ba74fcb0246b38a51b1578cf75be09905077750c19b1eb1\": rpc error: code = NotFound desc = could not find container \"9aeb281f96a9e4909ba74fcb0246b38a51b1578cf75be09905077750c19b1eb1\": container with ID starting with 9aeb281f96a9e4909ba74fcb0246b38a51b1578cf75be09905077750c19b1eb1 not found: ID does not exist" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.187680 4702 scope.go:117] "RemoveContainer" containerID="92fdee327707bb0c612cfd0011379e14b1ff8ee7af7c4eebc8db3c471a4eb0b9" Nov 25 13:50:38 crc kubenswrapper[4702]: E1125 13:50:38.192265 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92fdee327707bb0c612cfd0011379e14b1ff8ee7af7c4eebc8db3c471a4eb0b9\": container with ID starting with 92fdee327707bb0c612cfd0011379e14b1ff8ee7af7c4eebc8db3c471a4eb0b9 not found: ID does not exist" containerID="92fdee327707bb0c612cfd0011379e14b1ff8ee7af7c4eebc8db3c471a4eb0b9" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.192319 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92fdee327707bb0c612cfd0011379e14b1ff8ee7af7c4eebc8db3c471a4eb0b9"} err="failed to get container status \"92fdee327707bb0c612cfd0011379e14b1ff8ee7af7c4eebc8db3c471a4eb0b9\": rpc error: code = NotFound desc = could not find container \"92fdee327707bb0c612cfd0011379e14b1ff8ee7af7c4eebc8db3c471a4eb0b9\": container with ID starting with 92fdee327707bb0c612cfd0011379e14b1ff8ee7af7c4eebc8db3c471a4eb0b9 not found: ID does not exist" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.192349 4702 scope.go:117] "RemoveContainer" containerID="052bbacfe3923529d43f8faf0dcca91bd7695404ebd9b87324946c534e265603" Nov 25 13:50:38 crc kubenswrapper[4702]: E1125 13:50:38.192935 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"052bbacfe3923529d43f8faf0dcca91bd7695404ebd9b87324946c534e265603\": container with ID starting with 052bbacfe3923529d43f8faf0dcca91bd7695404ebd9b87324946c534e265603 not found: ID does not exist" containerID="052bbacfe3923529d43f8faf0dcca91bd7695404ebd9b87324946c534e265603" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.192996 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"052bbacfe3923529d43f8faf0dcca91bd7695404ebd9b87324946c534e265603"} err="failed to get container status \"052bbacfe3923529d43f8faf0dcca91bd7695404ebd9b87324946c534e265603\": rpc error: code = NotFound desc = could not find container \"052bbacfe3923529d43f8faf0dcca91bd7695404ebd9b87324946c534e265603\": container with ID starting with 052bbacfe3923529d43f8faf0dcca91bd7695404ebd9b87324946c534e265603 not found: ID does not exist" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.193052 4702 scope.go:117] "RemoveContainer" containerID="1ddbe9a5a1c75085ea60f719d36e9a81327d12f38b34f6072fd550416ba6c8f8" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.207385 4702 scope.go:117] "RemoveContainer" containerID="1ddbe9a5a1c75085ea60f719d36e9a81327d12f38b34f6072fd550416ba6c8f8" Nov 25 13:50:38 crc kubenswrapper[4702]: E1125 13:50:38.207835 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ddbe9a5a1c75085ea60f719d36e9a81327d12f38b34f6072fd550416ba6c8f8\": container with ID starting with 1ddbe9a5a1c75085ea60f719d36e9a81327d12f38b34f6072fd550416ba6c8f8 not found: ID does not exist" containerID="1ddbe9a5a1c75085ea60f719d36e9a81327d12f38b34f6072fd550416ba6c8f8" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.207913 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ddbe9a5a1c75085ea60f719d36e9a81327d12f38b34f6072fd550416ba6c8f8"} err="failed to get container status \"1ddbe9a5a1c75085ea60f719d36e9a81327d12f38b34f6072fd550416ba6c8f8\": rpc error: code = NotFound desc = could not find container \"1ddbe9a5a1c75085ea60f719d36e9a81327d12f38b34f6072fd550416ba6c8f8\": container with ID starting with 1ddbe9a5a1c75085ea60f719d36e9a81327d12f38b34f6072fd550416ba6c8f8 not found: ID does not exist" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.207949 4702 scope.go:117] "RemoveContainer" containerID="93047c8e45a295c8afe15bafe26622ba661947c901b3e30712219f6388c35216" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.222274 4702 scope.go:117] "RemoveContainer" containerID="5c8e883a9434d217bba8c7100e0a34f0c64b68b0a01d0f5520131ae3aaf78e12" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.238573 4702 scope.go:117] "RemoveContainer" containerID="ee2a8af9ab5feb0f3a51eb3cb536a2a8c229072a07438479512f4b80f5f7dcdb" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.242672 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68f8eda2-8222-4ecb-a2d0-316ed8398d58-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.252602 4702 scope.go:117] "RemoveContainer" containerID="93047c8e45a295c8afe15bafe26622ba661947c901b3e30712219f6388c35216" Nov 25 13:50:38 crc kubenswrapper[4702]: E1125 13:50:38.253061 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93047c8e45a295c8afe15bafe26622ba661947c901b3e30712219f6388c35216\": container with ID starting with 93047c8e45a295c8afe15bafe26622ba661947c901b3e30712219f6388c35216 not found: ID does not exist" containerID="93047c8e45a295c8afe15bafe26622ba661947c901b3e30712219f6388c35216" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.253109 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93047c8e45a295c8afe15bafe26622ba661947c901b3e30712219f6388c35216"} err="failed to get container status \"93047c8e45a295c8afe15bafe26622ba661947c901b3e30712219f6388c35216\": rpc error: code = NotFound desc = could not find container \"93047c8e45a295c8afe15bafe26622ba661947c901b3e30712219f6388c35216\": container with ID starting with 93047c8e45a295c8afe15bafe26622ba661947c901b3e30712219f6388c35216 not found: ID does not exist" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.253143 4702 scope.go:117] "RemoveContainer" containerID="5c8e883a9434d217bba8c7100e0a34f0c64b68b0a01d0f5520131ae3aaf78e12" Nov 25 13:50:38 crc kubenswrapper[4702]: E1125 13:50:38.253576 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c8e883a9434d217bba8c7100e0a34f0c64b68b0a01d0f5520131ae3aaf78e12\": container with ID starting with 5c8e883a9434d217bba8c7100e0a34f0c64b68b0a01d0f5520131ae3aaf78e12 not found: ID does not exist" containerID="5c8e883a9434d217bba8c7100e0a34f0c64b68b0a01d0f5520131ae3aaf78e12" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.253603 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c8e883a9434d217bba8c7100e0a34f0c64b68b0a01d0f5520131ae3aaf78e12"} err="failed to get container status \"5c8e883a9434d217bba8c7100e0a34f0c64b68b0a01d0f5520131ae3aaf78e12\": rpc error: code = NotFound desc = could not find container \"5c8e883a9434d217bba8c7100e0a34f0c64b68b0a01d0f5520131ae3aaf78e12\": container with ID starting with 5c8e883a9434d217bba8c7100e0a34f0c64b68b0a01d0f5520131ae3aaf78e12 not found: ID does not exist" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.253627 4702 scope.go:117] "RemoveContainer" containerID="ee2a8af9ab5feb0f3a51eb3cb536a2a8c229072a07438479512f4b80f5f7dcdb" Nov 25 13:50:38 crc kubenswrapper[4702]: E1125 13:50:38.254075 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee2a8af9ab5feb0f3a51eb3cb536a2a8c229072a07438479512f4b80f5f7dcdb\": container with ID starting with ee2a8af9ab5feb0f3a51eb3cb536a2a8c229072a07438479512f4b80f5f7dcdb not found: ID does not exist" containerID="ee2a8af9ab5feb0f3a51eb3cb536a2a8c229072a07438479512f4b80f5f7dcdb" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.254140 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee2a8af9ab5feb0f3a51eb3cb536a2a8c229072a07438479512f4b80f5f7dcdb"} err="failed to get container status \"ee2a8af9ab5feb0f3a51eb3cb536a2a8c229072a07438479512f4b80f5f7dcdb\": rpc error: code = NotFound desc = could not find container \"ee2a8af9ab5feb0f3a51eb3cb536a2a8c229072a07438479512f4b80f5f7dcdb\": container with ID starting with ee2a8af9ab5feb0f3a51eb3cb536a2a8c229072a07438479512f4b80f5f7dcdb not found: ID does not exist" Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.317890 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v9skf"] Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.328144 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-v9skf"] Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.345363 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jq6gh"] Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.354407 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jq6gh"] Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.360540 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-zqrgn"] Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.364825 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-zqrgn"] Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.369866 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ppzc7"] Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.374121 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ppzc7"] Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.385004 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pwlcj"] Nov 25 13:50:38 crc kubenswrapper[4702]: I1125 13:50:38.389872 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pwlcj"] Nov 25 13:50:39 crc kubenswrapper[4702]: I1125 13:50:39.007309 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8mh98" event={"ID":"b49d10e5-028c-4ad8-b03f-882744920b74","Type":"ContainerStarted","Data":"8267bcf8610d77bba8fa28c8f008f748e51fc99b3bfa86d32b9695c9850e5a48"} Nov 25 13:50:39 crc kubenswrapper[4702]: I1125 13:50:39.007389 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8mh98" event={"ID":"b49d10e5-028c-4ad8-b03f-882744920b74","Type":"ContainerStarted","Data":"79bd555370c3ca9cc8a9eb3d0e44780c9a6cc2468ff40a1d5895397afa911c53"} Nov 25 13:50:39 crc kubenswrapper[4702]: I1125 13:50:39.008001 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-8mh98" Nov 25 13:50:39 crc kubenswrapper[4702]: I1125 13:50:39.013079 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-8mh98" Nov 25 13:50:39 crc kubenswrapper[4702]: I1125 13:50:39.038885 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-8mh98" podStartSLOduration=2.038853255 podStartE2EDuration="2.038853255s" podCreationTimestamp="2025-11-25 13:50:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:50:39.036627776 +0000 UTC m=+248.199278296" watchObservedRunningTime="2025-11-25 13:50:39.038853255 +0000 UTC m=+248.201503715" Nov 25 13:50:39 crc kubenswrapper[4702]: I1125 13:50:39.810456 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0389edfa-9a86-48df-a9a6-26e2b4d092c4" path="/var/lib/kubelet/pods/0389edfa-9a86-48df-a9a6-26e2b4d092c4/volumes" Nov 25 13:50:39 crc kubenswrapper[4702]: I1125 13:50:39.812326 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="597c2ecd-d536-4eec-bc02-89c798cbf25c" path="/var/lib/kubelet/pods/597c2ecd-d536-4eec-bc02-89c798cbf25c/volumes" Nov 25 13:50:39 crc kubenswrapper[4702]: I1125 13:50:39.813325 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66c1bf9b-3a89-427b-9bba-009bdafcfcc8" path="/var/lib/kubelet/pods/66c1bf9b-3a89-427b-9bba-009bdafcfcc8/volumes" Nov 25 13:50:39 crc kubenswrapper[4702]: I1125 13:50:39.814922 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68f8eda2-8222-4ecb-a2d0-316ed8398d58" path="/var/lib/kubelet/pods/68f8eda2-8222-4ecb-a2d0-316ed8398d58/volumes" Nov 25 13:50:39 crc kubenswrapper[4702]: I1125 13:50:39.815808 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5b74e23-72c6-4278-9dc7-909f528d344b" path="/var/lib/kubelet/pods/f5b74e23-72c6-4278-9dc7-909f528d344b/volumes" Nov 25 13:50:47 crc kubenswrapper[4702]: I1125 13:50:47.536972 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rmtvx"] Nov 25 13:50:47 crc kubenswrapper[4702]: I1125 13:50:47.538165 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-rmtvx" podUID="5413dfbd-3c8b-42ac-8536-c0331fb7a397" containerName="controller-manager" containerID="cri-o://d5cab50337cc399fb7170add9ede78cdd42dea6d9c62d3d56a2c545af1c532e4" gracePeriod=30 Nov 25 13:50:47 crc kubenswrapper[4702]: I1125 13:50:47.640343 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wv6k6"] Nov 25 13:50:47 crc kubenswrapper[4702]: I1125 13:50:47.640714 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wv6k6" podUID="ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73" containerName="route-controller-manager" containerID="cri-o://6afeef1a8b74e4384f98678148072a6d7b5248820dda22d75b4a48b81b6fd22f" gracePeriod=30 Nov 25 13:50:47 crc kubenswrapper[4702]: I1125 13:50:47.910009 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rmtvx" Nov 25 13:50:47 crc kubenswrapper[4702]: I1125 13:50:47.985166 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wv6k6" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.071193 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wv6k6" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.071246 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wv6k6" event={"ID":"ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73","Type":"ContainerDied","Data":"6afeef1a8b74e4384f98678148072a6d7b5248820dda22d75b4a48b81b6fd22f"} Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.071303 4702 scope.go:117] "RemoveContainer" containerID="6afeef1a8b74e4384f98678148072a6d7b5248820dda22d75b4a48b81b6fd22f" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.071142 4702 generic.go:334] "Generic (PLEG): container finished" podID="ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73" containerID="6afeef1a8b74e4384f98678148072a6d7b5248820dda22d75b4a48b81b6fd22f" exitCode=0 Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.071459 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wv6k6" event={"ID":"ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73","Type":"ContainerDied","Data":"cdf2dd1f3ca165b58b9a99dbc1f7ad5e068a3d14b16ab3d05029a3402a409f4b"} Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.074713 4702 generic.go:334] "Generic (PLEG): container finished" podID="5413dfbd-3c8b-42ac-8536-c0331fb7a397" containerID="d5cab50337cc399fb7170add9ede78cdd42dea6d9c62d3d56a2c545af1c532e4" exitCode=0 Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.074750 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rmtvx" event={"ID":"5413dfbd-3c8b-42ac-8536-c0331fb7a397","Type":"ContainerDied","Data":"d5cab50337cc399fb7170add9ede78cdd42dea6d9c62d3d56a2c545af1c532e4"} Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.074772 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rmtvx" event={"ID":"5413dfbd-3c8b-42ac-8536-c0331fb7a397","Type":"ContainerDied","Data":"a78343cf17571cc0a3b9d85ea6f4740c2bde3132db48e84cd73f03f40258c4dd"} Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.074811 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rmtvx" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.088930 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5413dfbd-3c8b-42ac-8536-c0331fb7a397-config\") pod \"5413dfbd-3c8b-42ac-8536-c0331fb7a397\" (UID: \"5413dfbd-3c8b-42ac-8536-c0331fb7a397\") " Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.088979 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ccsm\" (UniqueName: \"kubernetes.io/projected/5413dfbd-3c8b-42ac-8536-c0331fb7a397-kube-api-access-7ccsm\") pod \"5413dfbd-3c8b-42ac-8536-c0331fb7a397\" (UID: \"5413dfbd-3c8b-42ac-8536-c0331fb7a397\") " Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.089030 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5413dfbd-3c8b-42ac-8536-c0331fb7a397-client-ca\") pod \"5413dfbd-3c8b-42ac-8536-c0331fb7a397\" (UID: \"5413dfbd-3c8b-42ac-8536-c0331fb7a397\") " Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.089149 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73-client-ca\") pod \"ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73\" (UID: \"ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73\") " Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.089215 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5413dfbd-3c8b-42ac-8536-c0331fb7a397-serving-cert\") pod \"5413dfbd-3c8b-42ac-8536-c0331fb7a397\" (UID: \"5413dfbd-3c8b-42ac-8536-c0331fb7a397\") " Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.089280 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73-serving-cert\") pod \"ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73\" (UID: \"ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73\") " Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.089311 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8cqf\" (UniqueName: \"kubernetes.io/projected/ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73-kube-api-access-q8cqf\") pod \"ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73\" (UID: \"ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73\") " Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.089342 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5413dfbd-3c8b-42ac-8536-c0331fb7a397-proxy-ca-bundles\") pod \"5413dfbd-3c8b-42ac-8536-c0331fb7a397\" (UID: \"5413dfbd-3c8b-42ac-8536-c0331fb7a397\") " Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.089377 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73-config\") pod \"ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73\" (UID: \"ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73\") " Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.090113 4702 scope.go:117] "RemoveContainer" containerID="6afeef1a8b74e4384f98678148072a6d7b5248820dda22d75b4a48b81b6fd22f" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.090892 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73-client-ca" (OuterVolumeSpecName: "client-ca") pod "ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73" (UID: "ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.091083 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5413dfbd-3c8b-42ac-8536-c0331fb7a397-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "5413dfbd-3c8b-42ac-8536-c0331fb7a397" (UID: "5413dfbd-3c8b-42ac-8536-c0331fb7a397"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.091138 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5413dfbd-3c8b-42ac-8536-c0331fb7a397-client-ca" (OuterVolumeSpecName: "client-ca") pod "5413dfbd-3c8b-42ac-8536-c0331fb7a397" (UID: "5413dfbd-3c8b-42ac-8536-c0331fb7a397"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.091602 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73-config" (OuterVolumeSpecName: "config") pod "ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73" (UID: "ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.091815 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5413dfbd-3c8b-42ac-8536-c0331fb7a397-config" (OuterVolumeSpecName: "config") pod "5413dfbd-3c8b-42ac-8536-c0331fb7a397" (UID: "5413dfbd-3c8b-42ac-8536-c0331fb7a397"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:50:48 crc kubenswrapper[4702]: E1125 13:50:48.093217 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6afeef1a8b74e4384f98678148072a6d7b5248820dda22d75b4a48b81b6fd22f\": container with ID starting with 6afeef1a8b74e4384f98678148072a6d7b5248820dda22d75b4a48b81b6fd22f not found: ID does not exist" containerID="6afeef1a8b74e4384f98678148072a6d7b5248820dda22d75b4a48b81b6fd22f" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.093354 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6afeef1a8b74e4384f98678148072a6d7b5248820dda22d75b4a48b81b6fd22f"} err="failed to get container status \"6afeef1a8b74e4384f98678148072a6d7b5248820dda22d75b4a48b81b6fd22f\": rpc error: code = NotFound desc = could not find container \"6afeef1a8b74e4384f98678148072a6d7b5248820dda22d75b4a48b81b6fd22f\": container with ID starting with 6afeef1a8b74e4384f98678148072a6d7b5248820dda22d75b4a48b81b6fd22f not found: ID does not exist" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.093391 4702 scope.go:117] "RemoveContainer" containerID="d5cab50337cc399fb7170add9ede78cdd42dea6d9c62d3d56a2c545af1c532e4" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.098094 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5413dfbd-3c8b-42ac-8536-c0331fb7a397-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5413dfbd-3c8b-42ac-8536-c0331fb7a397" (UID: "5413dfbd-3c8b-42ac-8536-c0331fb7a397"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.098282 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5413dfbd-3c8b-42ac-8536-c0331fb7a397-kube-api-access-7ccsm" (OuterVolumeSpecName: "kube-api-access-7ccsm") pod "5413dfbd-3c8b-42ac-8536-c0331fb7a397" (UID: "5413dfbd-3c8b-42ac-8536-c0331fb7a397"). InnerVolumeSpecName "kube-api-access-7ccsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.098316 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73-kube-api-access-q8cqf" (OuterVolumeSpecName: "kube-api-access-q8cqf") pod "ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73" (UID: "ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73"). InnerVolumeSpecName "kube-api-access-q8cqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.098638 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73" (UID: "ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.111728 4702 scope.go:117] "RemoveContainer" containerID="d5cab50337cc399fb7170add9ede78cdd42dea6d9c62d3d56a2c545af1c532e4" Nov 25 13:50:48 crc kubenswrapper[4702]: E1125 13:50:48.112187 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5cab50337cc399fb7170add9ede78cdd42dea6d9c62d3d56a2c545af1c532e4\": container with ID starting with d5cab50337cc399fb7170add9ede78cdd42dea6d9c62d3d56a2c545af1c532e4 not found: ID does not exist" containerID="d5cab50337cc399fb7170add9ede78cdd42dea6d9c62d3d56a2c545af1c532e4" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.112257 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5cab50337cc399fb7170add9ede78cdd42dea6d9c62d3d56a2c545af1c532e4"} err="failed to get container status \"d5cab50337cc399fb7170add9ede78cdd42dea6d9c62d3d56a2c545af1c532e4\": rpc error: code = NotFound desc = could not find container \"d5cab50337cc399fb7170add9ede78cdd42dea6d9c62d3d56a2c545af1c532e4\": container with ID starting with d5cab50337cc399fb7170add9ede78cdd42dea6d9c62d3d56a2c545af1c532e4 not found: ID does not exist" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.191703 4702 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.191752 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5413dfbd-3c8b-42ac-8536-c0331fb7a397-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.191768 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.191780 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8cqf\" (UniqueName: \"kubernetes.io/projected/ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73-kube-api-access-q8cqf\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.191797 4702 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5413dfbd-3c8b-42ac-8536-c0331fb7a397-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.191808 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73-config\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.191819 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5413dfbd-3c8b-42ac-8536-c0331fb7a397-config\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.191831 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ccsm\" (UniqueName: \"kubernetes.io/projected/5413dfbd-3c8b-42ac-8536-c0331fb7a397-kube-api-access-7ccsm\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.191842 4702 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5413dfbd-3c8b-42ac-8536-c0331fb7a397-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.399329 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wv6k6"] Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.407729 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wv6k6"] Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.411636 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rmtvx"] Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.415438 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rmtvx"] Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.836351 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-88759c5b4-xjptl"] Nov 25 13:50:48 crc kubenswrapper[4702]: E1125 13:50:48.836632 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5b74e23-72c6-4278-9dc7-909f528d344b" containerName="extract-utilities" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.836649 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5b74e23-72c6-4278-9dc7-909f528d344b" containerName="extract-utilities" Nov 25 13:50:48 crc kubenswrapper[4702]: E1125 13:50:48.836660 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="597c2ecd-d536-4eec-bc02-89c798cbf25c" containerName="extract-content" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.836667 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="597c2ecd-d536-4eec-bc02-89c798cbf25c" containerName="extract-content" Nov 25 13:50:48 crc kubenswrapper[4702]: E1125 13:50:48.836692 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66c1bf9b-3a89-427b-9bba-009bdafcfcc8" containerName="extract-utilities" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.836699 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="66c1bf9b-3a89-427b-9bba-009bdafcfcc8" containerName="extract-utilities" Nov 25 13:50:48 crc kubenswrapper[4702]: E1125 13:50:48.836709 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0389edfa-9a86-48df-a9a6-26e2b4d092c4" containerName="marketplace-operator" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.836715 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="0389edfa-9a86-48df-a9a6-26e2b4d092c4" containerName="marketplace-operator" Nov 25 13:50:48 crc kubenswrapper[4702]: E1125 13:50:48.836722 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66c1bf9b-3a89-427b-9bba-009bdafcfcc8" containerName="extract-content" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.836730 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="66c1bf9b-3a89-427b-9bba-009bdafcfcc8" containerName="extract-content" Nov 25 13:50:48 crc kubenswrapper[4702]: E1125 13:50:48.836739 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="597c2ecd-d536-4eec-bc02-89c798cbf25c" containerName="extract-utilities" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.836746 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="597c2ecd-d536-4eec-bc02-89c798cbf25c" containerName="extract-utilities" Nov 25 13:50:48 crc kubenswrapper[4702]: E1125 13:50:48.836754 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68f8eda2-8222-4ecb-a2d0-316ed8398d58" containerName="extract-content" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.836761 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="68f8eda2-8222-4ecb-a2d0-316ed8398d58" containerName="extract-content" Nov 25 13:50:48 crc kubenswrapper[4702]: E1125 13:50:48.836771 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73" containerName="route-controller-manager" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.836778 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73" containerName="route-controller-manager" Nov 25 13:50:48 crc kubenswrapper[4702]: E1125 13:50:48.836787 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68f8eda2-8222-4ecb-a2d0-316ed8398d58" containerName="registry-server" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.836794 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="68f8eda2-8222-4ecb-a2d0-316ed8398d58" containerName="registry-server" Nov 25 13:50:48 crc kubenswrapper[4702]: E1125 13:50:48.836802 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5b74e23-72c6-4278-9dc7-909f528d344b" containerName="registry-server" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.836809 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5b74e23-72c6-4278-9dc7-909f528d344b" containerName="registry-server" Nov 25 13:50:48 crc kubenswrapper[4702]: E1125 13:50:48.836817 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="597c2ecd-d536-4eec-bc02-89c798cbf25c" containerName="registry-server" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.836823 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="597c2ecd-d536-4eec-bc02-89c798cbf25c" containerName="registry-server" Nov 25 13:50:48 crc kubenswrapper[4702]: E1125 13:50:48.836830 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66c1bf9b-3a89-427b-9bba-009bdafcfcc8" containerName="registry-server" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.836835 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="66c1bf9b-3a89-427b-9bba-009bdafcfcc8" containerName="registry-server" Nov 25 13:50:48 crc kubenswrapper[4702]: E1125 13:50:48.836847 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68f8eda2-8222-4ecb-a2d0-316ed8398d58" containerName="extract-utilities" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.836853 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="68f8eda2-8222-4ecb-a2d0-316ed8398d58" containerName="extract-utilities" Nov 25 13:50:48 crc kubenswrapper[4702]: E1125 13:50:48.836860 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5413dfbd-3c8b-42ac-8536-c0331fb7a397" containerName="controller-manager" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.836866 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="5413dfbd-3c8b-42ac-8536-c0331fb7a397" containerName="controller-manager" Nov 25 13:50:48 crc kubenswrapper[4702]: E1125 13:50:48.836875 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5b74e23-72c6-4278-9dc7-909f528d344b" containerName="extract-content" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.836881 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5b74e23-72c6-4278-9dc7-909f528d344b" containerName="extract-content" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.836974 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="5413dfbd-3c8b-42ac-8536-c0331fb7a397" containerName="controller-manager" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.836984 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5b74e23-72c6-4278-9dc7-909f528d344b" containerName="registry-server" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.836993 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="597c2ecd-d536-4eec-bc02-89c798cbf25c" containerName="registry-server" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.837002 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73" containerName="route-controller-manager" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.837013 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="68f8eda2-8222-4ecb-a2d0-316ed8398d58" containerName="registry-server" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.837020 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="0389edfa-9a86-48df-a9a6-26e2b4d092c4" containerName="marketplace-operator" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.837029 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="66c1bf9b-3a89-427b-9bba-009bdafcfcc8" containerName="registry-server" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.837458 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-88759c5b4-xjptl" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.839824 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.840198 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.840196 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.840525 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.840553 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.840588 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.846359 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c5df878b4-4kk5g"] Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.847535 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c5df878b4-4kk5g" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.848455 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.850772 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.850821 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.851012 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.851133 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.851255 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.851521 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.852853 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c5df878b4-4kk5g"] Nov 25 13:50:48 crc kubenswrapper[4702]: I1125 13:50:48.856849 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-88759c5b4-xjptl"] Nov 25 13:50:49 crc kubenswrapper[4702]: I1125 13:50:49.004028 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/97421c1d-3038-4993-8a2b-71b7629384fe-client-ca\") pod \"route-controller-manager-5c5df878b4-4kk5g\" (UID: \"97421c1d-3038-4993-8a2b-71b7629384fe\") " pod="openshift-route-controller-manager/route-controller-manager-5c5df878b4-4kk5g" Nov 25 13:50:49 crc kubenswrapper[4702]: I1125 13:50:49.004105 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/84464ec3-791f-4068-8b21-e3afbde59e97-proxy-ca-bundles\") pod \"controller-manager-88759c5b4-xjptl\" (UID: \"84464ec3-791f-4068-8b21-e3afbde59e97\") " pod="openshift-controller-manager/controller-manager-88759c5b4-xjptl" Nov 25 13:50:49 crc kubenswrapper[4702]: I1125 13:50:49.004313 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84464ec3-791f-4068-8b21-e3afbde59e97-config\") pod \"controller-manager-88759c5b4-xjptl\" (UID: \"84464ec3-791f-4068-8b21-e3afbde59e97\") " pod="openshift-controller-manager/controller-manager-88759c5b4-xjptl" Nov 25 13:50:49 crc kubenswrapper[4702]: I1125 13:50:49.004467 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84464ec3-791f-4068-8b21-e3afbde59e97-serving-cert\") pod \"controller-manager-88759c5b4-xjptl\" (UID: \"84464ec3-791f-4068-8b21-e3afbde59e97\") " pod="openshift-controller-manager/controller-manager-88759c5b4-xjptl" Nov 25 13:50:49 crc kubenswrapper[4702]: I1125 13:50:49.004575 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9f9pq\" (UniqueName: \"kubernetes.io/projected/84464ec3-791f-4068-8b21-e3afbde59e97-kube-api-access-9f9pq\") pod \"controller-manager-88759c5b4-xjptl\" (UID: \"84464ec3-791f-4068-8b21-e3afbde59e97\") " pod="openshift-controller-manager/controller-manager-88759c5b4-xjptl" Nov 25 13:50:49 crc kubenswrapper[4702]: I1125 13:50:49.004611 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97421c1d-3038-4993-8a2b-71b7629384fe-config\") pod \"route-controller-manager-5c5df878b4-4kk5g\" (UID: \"97421c1d-3038-4993-8a2b-71b7629384fe\") " pod="openshift-route-controller-manager/route-controller-manager-5c5df878b4-4kk5g" Nov 25 13:50:49 crc kubenswrapper[4702]: I1125 13:50:49.004721 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97421c1d-3038-4993-8a2b-71b7629384fe-serving-cert\") pod \"route-controller-manager-5c5df878b4-4kk5g\" (UID: \"97421c1d-3038-4993-8a2b-71b7629384fe\") " pod="openshift-route-controller-manager/route-controller-manager-5c5df878b4-4kk5g" Nov 25 13:50:49 crc kubenswrapper[4702]: I1125 13:50:49.004818 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/84464ec3-791f-4068-8b21-e3afbde59e97-client-ca\") pod \"controller-manager-88759c5b4-xjptl\" (UID: \"84464ec3-791f-4068-8b21-e3afbde59e97\") " pod="openshift-controller-manager/controller-manager-88759c5b4-xjptl" Nov 25 13:50:49 crc kubenswrapper[4702]: I1125 13:50:49.004881 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l67zj\" (UniqueName: \"kubernetes.io/projected/97421c1d-3038-4993-8a2b-71b7629384fe-kube-api-access-l67zj\") pod \"route-controller-manager-5c5df878b4-4kk5g\" (UID: \"97421c1d-3038-4993-8a2b-71b7629384fe\") " pod="openshift-route-controller-manager/route-controller-manager-5c5df878b4-4kk5g" Nov 25 13:50:49 crc kubenswrapper[4702]: I1125 13:50:49.106323 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97421c1d-3038-4993-8a2b-71b7629384fe-serving-cert\") pod \"route-controller-manager-5c5df878b4-4kk5g\" (UID: \"97421c1d-3038-4993-8a2b-71b7629384fe\") " pod="openshift-route-controller-manager/route-controller-manager-5c5df878b4-4kk5g" Nov 25 13:50:49 crc kubenswrapper[4702]: I1125 13:50:49.106378 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/84464ec3-791f-4068-8b21-e3afbde59e97-client-ca\") pod \"controller-manager-88759c5b4-xjptl\" (UID: \"84464ec3-791f-4068-8b21-e3afbde59e97\") " pod="openshift-controller-manager/controller-manager-88759c5b4-xjptl" Nov 25 13:50:49 crc kubenswrapper[4702]: I1125 13:50:49.106418 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l67zj\" (UniqueName: \"kubernetes.io/projected/97421c1d-3038-4993-8a2b-71b7629384fe-kube-api-access-l67zj\") pod \"route-controller-manager-5c5df878b4-4kk5g\" (UID: \"97421c1d-3038-4993-8a2b-71b7629384fe\") " pod="openshift-route-controller-manager/route-controller-manager-5c5df878b4-4kk5g" Nov 25 13:50:49 crc kubenswrapper[4702]: I1125 13:50:49.106443 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/97421c1d-3038-4993-8a2b-71b7629384fe-client-ca\") pod \"route-controller-manager-5c5df878b4-4kk5g\" (UID: \"97421c1d-3038-4993-8a2b-71b7629384fe\") " pod="openshift-route-controller-manager/route-controller-manager-5c5df878b4-4kk5g" Nov 25 13:50:49 crc kubenswrapper[4702]: I1125 13:50:49.106469 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/84464ec3-791f-4068-8b21-e3afbde59e97-proxy-ca-bundles\") pod \"controller-manager-88759c5b4-xjptl\" (UID: \"84464ec3-791f-4068-8b21-e3afbde59e97\") " pod="openshift-controller-manager/controller-manager-88759c5b4-xjptl" Nov 25 13:50:49 crc kubenswrapper[4702]: I1125 13:50:49.106498 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84464ec3-791f-4068-8b21-e3afbde59e97-config\") pod \"controller-manager-88759c5b4-xjptl\" (UID: \"84464ec3-791f-4068-8b21-e3afbde59e97\") " pod="openshift-controller-manager/controller-manager-88759c5b4-xjptl" Nov 25 13:50:49 crc kubenswrapper[4702]: I1125 13:50:49.106526 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84464ec3-791f-4068-8b21-e3afbde59e97-serving-cert\") pod \"controller-manager-88759c5b4-xjptl\" (UID: \"84464ec3-791f-4068-8b21-e3afbde59e97\") " pod="openshift-controller-manager/controller-manager-88759c5b4-xjptl" Nov 25 13:50:49 crc kubenswrapper[4702]: I1125 13:50:49.106558 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9f9pq\" (UniqueName: \"kubernetes.io/projected/84464ec3-791f-4068-8b21-e3afbde59e97-kube-api-access-9f9pq\") pod \"controller-manager-88759c5b4-xjptl\" (UID: \"84464ec3-791f-4068-8b21-e3afbde59e97\") " pod="openshift-controller-manager/controller-manager-88759c5b4-xjptl" Nov 25 13:50:49 crc kubenswrapper[4702]: I1125 13:50:49.106578 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97421c1d-3038-4993-8a2b-71b7629384fe-config\") pod \"route-controller-manager-5c5df878b4-4kk5g\" (UID: \"97421c1d-3038-4993-8a2b-71b7629384fe\") " pod="openshift-route-controller-manager/route-controller-manager-5c5df878b4-4kk5g" Nov 25 13:50:49 crc kubenswrapper[4702]: I1125 13:50:49.107750 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/84464ec3-791f-4068-8b21-e3afbde59e97-client-ca\") pod \"controller-manager-88759c5b4-xjptl\" (UID: \"84464ec3-791f-4068-8b21-e3afbde59e97\") " pod="openshift-controller-manager/controller-manager-88759c5b4-xjptl" Nov 25 13:50:49 crc kubenswrapper[4702]: I1125 13:50:49.107960 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/97421c1d-3038-4993-8a2b-71b7629384fe-client-ca\") pod \"route-controller-manager-5c5df878b4-4kk5g\" (UID: \"97421c1d-3038-4993-8a2b-71b7629384fe\") " pod="openshift-route-controller-manager/route-controller-manager-5c5df878b4-4kk5g" Nov 25 13:50:49 crc kubenswrapper[4702]: I1125 13:50:49.108175 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97421c1d-3038-4993-8a2b-71b7629384fe-config\") pod \"route-controller-manager-5c5df878b4-4kk5g\" (UID: \"97421c1d-3038-4993-8a2b-71b7629384fe\") " pod="openshift-route-controller-manager/route-controller-manager-5c5df878b4-4kk5g" Nov 25 13:50:49 crc kubenswrapper[4702]: I1125 13:50:49.108204 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/84464ec3-791f-4068-8b21-e3afbde59e97-proxy-ca-bundles\") pod \"controller-manager-88759c5b4-xjptl\" (UID: \"84464ec3-791f-4068-8b21-e3afbde59e97\") " pod="openshift-controller-manager/controller-manager-88759c5b4-xjptl" Nov 25 13:50:49 crc kubenswrapper[4702]: I1125 13:50:49.108879 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84464ec3-791f-4068-8b21-e3afbde59e97-config\") pod \"controller-manager-88759c5b4-xjptl\" (UID: \"84464ec3-791f-4068-8b21-e3afbde59e97\") " pod="openshift-controller-manager/controller-manager-88759c5b4-xjptl" Nov 25 13:50:49 crc kubenswrapper[4702]: I1125 13:50:49.113693 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97421c1d-3038-4993-8a2b-71b7629384fe-serving-cert\") pod \"route-controller-manager-5c5df878b4-4kk5g\" (UID: \"97421c1d-3038-4993-8a2b-71b7629384fe\") " pod="openshift-route-controller-manager/route-controller-manager-5c5df878b4-4kk5g" Nov 25 13:50:49 crc kubenswrapper[4702]: I1125 13:50:49.114025 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84464ec3-791f-4068-8b21-e3afbde59e97-serving-cert\") pod \"controller-manager-88759c5b4-xjptl\" (UID: \"84464ec3-791f-4068-8b21-e3afbde59e97\") " pod="openshift-controller-manager/controller-manager-88759c5b4-xjptl" Nov 25 13:50:49 crc kubenswrapper[4702]: I1125 13:50:49.125779 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9f9pq\" (UniqueName: \"kubernetes.io/projected/84464ec3-791f-4068-8b21-e3afbde59e97-kube-api-access-9f9pq\") pod \"controller-manager-88759c5b4-xjptl\" (UID: \"84464ec3-791f-4068-8b21-e3afbde59e97\") " pod="openshift-controller-manager/controller-manager-88759c5b4-xjptl" Nov 25 13:50:49 crc kubenswrapper[4702]: I1125 13:50:49.127422 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l67zj\" (UniqueName: \"kubernetes.io/projected/97421c1d-3038-4993-8a2b-71b7629384fe-kube-api-access-l67zj\") pod \"route-controller-manager-5c5df878b4-4kk5g\" (UID: \"97421c1d-3038-4993-8a2b-71b7629384fe\") " pod="openshift-route-controller-manager/route-controller-manager-5c5df878b4-4kk5g" Nov 25 13:50:49 crc kubenswrapper[4702]: I1125 13:50:49.162509 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-88759c5b4-xjptl" Nov 25 13:50:49 crc kubenswrapper[4702]: I1125 13:50:49.174961 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c5df878b4-4kk5g" Nov 25 13:50:49 crc kubenswrapper[4702]: I1125 13:50:49.394333 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-88759c5b4-xjptl"] Nov 25 13:50:49 crc kubenswrapper[4702]: W1125 13:50:49.403939 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod84464ec3_791f_4068_8b21_e3afbde59e97.slice/crio-78d05266ecf34af9f35184f384e62552a0ee898a906b1a69e5741a208c45f3b8 WatchSource:0}: Error finding container 78d05266ecf34af9f35184f384e62552a0ee898a906b1a69e5741a208c45f3b8: Status 404 returned error can't find the container with id 78d05266ecf34af9f35184f384e62552a0ee898a906b1a69e5741a208c45f3b8 Nov 25 13:50:49 crc kubenswrapper[4702]: I1125 13:50:49.435135 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c5df878b4-4kk5g"] Nov 25 13:50:49 crc kubenswrapper[4702]: W1125 13:50:49.438658 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97421c1d_3038_4993_8a2b_71b7629384fe.slice/crio-26a433607349f18a8a4a6a9093011d8ef06d1807b1d01e52308c30ae075058a4 WatchSource:0}: Error finding container 26a433607349f18a8a4a6a9093011d8ef06d1807b1d01e52308c30ae075058a4: Status 404 returned error can't find the container with id 26a433607349f18a8a4a6a9093011d8ef06d1807b1d01e52308c30ae075058a4 Nov 25 13:50:49 crc kubenswrapper[4702]: I1125 13:50:49.808847 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5413dfbd-3c8b-42ac-8536-c0331fb7a397" path="/var/lib/kubelet/pods/5413dfbd-3c8b-42ac-8536-c0331fb7a397/volumes" Nov 25 13:50:49 crc kubenswrapper[4702]: I1125 13:50:49.810191 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73" path="/var/lib/kubelet/pods/ea52b42e-7b7d-4f35-9d6d-9a7f46a95f73/volumes" Nov 25 13:50:50 crc kubenswrapper[4702]: I1125 13:50:50.091452 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c5df878b4-4kk5g" event={"ID":"97421c1d-3038-4993-8a2b-71b7629384fe","Type":"ContainerStarted","Data":"43265382df7d88cf229f78523838d46834521f8a7fddee8dc8115da420145574"} Nov 25 13:50:50 crc kubenswrapper[4702]: I1125 13:50:50.091535 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c5df878b4-4kk5g" event={"ID":"97421c1d-3038-4993-8a2b-71b7629384fe","Type":"ContainerStarted","Data":"26a433607349f18a8a4a6a9093011d8ef06d1807b1d01e52308c30ae075058a4"} Nov 25 13:50:50 crc kubenswrapper[4702]: I1125 13:50:50.091962 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5c5df878b4-4kk5g" Nov 25 13:50:50 crc kubenswrapper[4702]: I1125 13:50:50.093138 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-88759c5b4-xjptl" event={"ID":"84464ec3-791f-4068-8b21-e3afbde59e97","Type":"ContainerStarted","Data":"94e7995139f29df1142a77015b54ddad293217bb450c398c99a266a8947f35a6"} Nov 25 13:50:50 crc kubenswrapper[4702]: I1125 13:50:50.093170 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-88759c5b4-xjptl" event={"ID":"84464ec3-791f-4068-8b21-e3afbde59e97","Type":"ContainerStarted","Data":"78d05266ecf34af9f35184f384e62552a0ee898a906b1a69e5741a208c45f3b8"} Nov 25 13:50:50 crc kubenswrapper[4702]: I1125 13:50:50.093834 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-88759c5b4-xjptl" Nov 25 13:50:50 crc kubenswrapper[4702]: I1125 13:50:50.099537 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-88759c5b4-xjptl" Nov 25 13:50:50 crc kubenswrapper[4702]: I1125 13:50:50.099607 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5c5df878b4-4kk5g" Nov 25 13:50:50 crc kubenswrapper[4702]: I1125 13:50:50.124771 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5c5df878b4-4kk5g" podStartSLOduration=3.124745785 podStartE2EDuration="3.124745785s" podCreationTimestamp="2025-11-25 13:50:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:50:50.120132874 +0000 UTC m=+259.282783324" watchObservedRunningTime="2025-11-25 13:50:50.124745785 +0000 UTC m=+259.287396235" Nov 25 13:50:50 crc kubenswrapper[4702]: I1125 13:50:50.154488 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-88759c5b4-xjptl" podStartSLOduration=3.154470753 podStartE2EDuration="3.154470753s" podCreationTimestamp="2025-11-25 13:50:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:50:50.151769193 +0000 UTC m=+259.314419643" watchObservedRunningTime="2025-11-25 13:50:50.154470753 +0000 UTC m=+259.317121203" Nov 25 13:50:53 crc kubenswrapper[4702]: I1125 13:50:53.842732 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-88759c5b4-xjptl"] Nov 25 13:50:53 crc kubenswrapper[4702]: I1125 13:50:53.843275 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-88759c5b4-xjptl" podUID="84464ec3-791f-4068-8b21-e3afbde59e97" containerName="controller-manager" containerID="cri-o://94e7995139f29df1142a77015b54ddad293217bb450c398c99a266a8947f35a6" gracePeriod=30 Nov 25 13:50:53 crc kubenswrapper[4702]: I1125 13:50:53.861771 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c5df878b4-4kk5g"] Nov 25 13:50:53 crc kubenswrapper[4702]: I1125 13:50:53.862489 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5c5df878b4-4kk5g" podUID="97421c1d-3038-4993-8a2b-71b7629384fe" containerName="route-controller-manager" containerID="cri-o://43265382df7d88cf229f78523838d46834521f8a7fddee8dc8115da420145574" gracePeriod=30 Nov 25 13:50:54 crc kubenswrapper[4702]: I1125 13:50:54.117727 4702 generic.go:334] "Generic (PLEG): container finished" podID="84464ec3-791f-4068-8b21-e3afbde59e97" containerID="94e7995139f29df1142a77015b54ddad293217bb450c398c99a266a8947f35a6" exitCode=0 Nov 25 13:50:54 crc kubenswrapper[4702]: I1125 13:50:54.117811 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-88759c5b4-xjptl" event={"ID":"84464ec3-791f-4068-8b21-e3afbde59e97","Type":"ContainerDied","Data":"94e7995139f29df1142a77015b54ddad293217bb450c398c99a266a8947f35a6"} Nov 25 13:50:54 crc kubenswrapper[4702]: I1125 13:50:54.120519 4702 generic.go:334] "Generic (PLEG): container finished" podID="97421c1d-3038-4993-8a2b-71b7629384fe" containerID="43265382df7d88cf229f78523838d46834521f8a7fddee8dc8115da420145574" exitCode=0 Nov 25 13:50:54 crc kubenswrapper[4702]: I1125 13:50:54.120607 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c5df878b4-4kk5g" event={"ID":"97421c1d-3038-4993-8a2b-71b7629384fe","Type":"ContainerDied","Data":"43265382df7d88cf229f78523838d46834521f8a7fddee8dc8115da420145574"} Nov 25 13:50:54 crc kubenswrapper[4702]: I1125 13:50:54.266504 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c5df878b4-4kk5g" Nov 25 13:50:54 crc kubenswrapper[4702]: I1125 13:50:54.337354 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-88759c5b4-xjptl" Nov 25 13:50:54 crc kubenswrapper[4702]: I1125 13:50:54.375623 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97421c1d-3038-4993-8a2b-71b7629384fe-config\") pod \"97421c1d-3038-4993-8a2b-71b7629384fe\" (UID: \"97421c1d-3038-4993-8a2b-71b7629384fe\") " Nov 25 13:50:54 crc kubenswrapper[4702]: I1125 13:50:54.375992 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97421c1d-3038-4993-8a2b-71b7629384fe-serving-cert\") pod \"97421c1d-3038-4993-8a2b-71b7629384fe\" (UID: \"97421c1d-3038-4993-8a2b-71b7629384fe\") " Nov 25 13:50:54 crc kubenswrapper[4702]: I1125 13:50:54.376088 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l67zj\" (UniqueName: \"kubernetes.io/projected/97421c1d-3038-4993-8a2b-71b7629384fe-kube-api-access-l67zj\") pod \"97421c1d-3038-4993-8a2b-71b7629384fe\" (UID: \"97421c1d-3038-4993-8a2b-71b7629384fe\") " Nov 25 13:50:54 crc kubenswrapper[4702]: I1125 13:50:54.376174 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/97421c1d-3038-4993-8a2b-71b7629384fe-client-ca\") pod \"97421c1d-3038-4993-8a2b-71b7629384fe\" (UID: \"97421c1d-3038-4993-8a2b-71b7629384fe\") " Nov 25 13:50:54 crc kubenswrapper[4702]: I1125 13:50:54.376309 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97421c1d-3038-4993-8a2b-71b7629384fe-config" (OuterVolumeSpecName: "config") pod "97421c1d-3038-4993-8a2b-71b7629384fe" (UID: "97421c1d-3038-4993-8a2b-71b7629384fe"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:50:54 crc kubenswrapper[4702]: I1125 13:50:54.376567 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97421c1d-3038-4993-8a2b-71b7629384fe-config\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:54 crc kubenswrapper[4702]: I1125 13:50:54.376611 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97421c1d-3038-4993-8a2b-71b7629384fe-client-ca" (OuterVolumeSpecName: "client-ca") pod "97421c1d-3038-4993-8a2b-71b7629384fe" (UID: "97421c1d-3038-4993-8a2b-71b7629384fe"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:50:54 crc kubenswrapper[4702]: I1125 13:50:54.381488 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97421c1d-3038-4993-8a2b-71b7629384fe-kube-api-access-l67zj" (OuterVolumeSpecName: "kube-api-access-l67zj") pod "97421c1d-3038-4993-8a2b-71b7629384fe" (UID: "97421c1d-3038-4993-8a2b-71b7629384fe"). InnerVolumeSpecName "kube-api-access-l67zj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:50:54 crc kubenswrapper[4702]: I1125 13:50:54.381558 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97421c1d-3038-4993-8a2b-71b7629384fe-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "97421c1d-3038-4993-8a2b-71b7629384fe" (UID: "97421c1d-3038-4993-8a2b-71b7629384fe"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:50:54 crc kubenswrapper[4702]: I1125 13:50:54.477724 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9f9pq\" (UniqueName: \"kubernetes.io/projected/84464ec3-791f-4068-8b21-e3afbde59e97-kube-api-access-9f9pq\") pod \"84464ec3-791f-4068-8b21-e3afbde59e97\" (UID: \"84464ec3-791f-4068-8b21-e3afbde59e97\") " Nov 25 13:50:54 crc kubenswrapper[4702]: I1125 13:50:54.477820 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/84464ec3-791f-4068-8b21-e3afbde59e97-proxy-ca-bundles\") pod \"84464ec3-791f-4068-8b21-e3afbde59e97\" (UID: \"84464ec3-791f-4068-8b21-e3afbde59e97\") " Nov 25 13:50:54 crc kubenswrapper[4702]: I1125 13:50:54.477901 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84464ec3-791f-4068-8b21-e3afbde59e97-serving-cert\") pod \"84464ec3-791f-4068-8b21-e3afbde59e97\" (UID: \"84464ec3-791f-4068-8b21-e3afbde59e97\") " Nov 25 13:50:54 crc kubenswrapper[4702]: I1125 13:50:54.477938 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84464ec3-791f-4068-8b21-e3afbde59e97-config\") pod \"84464ec3-791f-4068-8b21-e3afbde59e97\" (UID: \"84464ec3-791f-4068-8b21-e3afbde59e97\") " Nov 25 13:50:54 crc kubenswrapper[4702]: I1125 13:50:54.477958 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/84464ec3-791f-4068-8b21-e3afbde59e97-client-ca\") pod \"84464ec3-791f-4068-8b21-e3afbde59e97\" (UID: \"84464ec3-791f-4068-8b21-e3afbde59e97\") " Nov 25 13:50:54 crc kubenswrapper[4702]: I1125 13:50:54.478155 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97421c1d-3038-4993-8a2b-71b7629384fe-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:54 crc kubenswrapper[4702]: I1125 13:50:54.478167 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l67zj\" (UniqueName: \"kubernetes.io/projected/97421c1d-3038-4993-8a2b-71b7629384fe-kube-api-access-l67zj\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:54 crc kubenswrapper[4702]: I1125 13:50:54.478177 4702 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/97421c1d-3038-4993-8a2b-71b7629384fe-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:54 crc kubenswrapper[4702]: I1125 13:50:54.478689 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84464ec3-791f-4068-8b21-e3afbde59e97-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "84464ec3-791f-4068-8b21-e3afbde59e97" (UID: "84464ec3-791f-4068-8b21-e3afbde59e97"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:50:54 crc kubenswrapper[4702]: I1125 13:50:54.478747 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84464ec3-791f-4068-8b21-e3afbde59e97-client-ca" (OuterVolumeSpecName: "client-ca") pod "84464ec3-791f-4068-8b21-e3afbde59e97" (UID: "84464ec3-791f-4068-8b21-e3afbde59e97"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:50:54 crc kubenswrapper[4702]: I1125 13:50:54.479130 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84464ec3-791f-4068-8b21-e3afbde59e97-config" (OuterVolumeSpecName: "config") pod "84464ec3-791f-4068-8b21-e3afbde59e97" (UID: "84464ec3-791f-4068-8b21-e3afbde59e97"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:50:54 crc kubenswrapper[4702]: I1125 13:50:54.480634 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84464ec3-791f-4068-8b21-e3afbde59e97-kube-api-access-9f9pq" (OuterVolumeSpecName: "kube-api-access-9f9pq") pod "84464ec3-791f-4068-8b21-e3afbde59e97" (UID: "84464ec3-791f-4068-8b21-e3afbde59e97"). InnerVolumeSpecName "kube-api-access-9f9pq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:50:54 crc kubenswrapper[4702]: I1125 13:50:54.481203 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84464ec3-791f-4068-8b21-e3afbde59e97-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "84464ec3-791f-4068-8b21-e3afbde59e97" (UID: "84464ec3-791f-4068-8b21-e3afbde59e97"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:50:54 crc kubenswrapper[4702]: I1125 13:50:54.579666 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9f9pq\" (UniqueName: \"kubernetes.io/projected/84464ec3-791f-4068-8b21-e3afbde59e97-kube-api-access-9f9pq\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:54 crc kubenswrapper[4702]: I1125 13:50:54.579725 4702 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/84464ec3-791f-4068-8b21-e3afbde59e97-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:54 crc kubenswrapper[4702]: I1125 13:50:54.579738 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84464ec3-791f-4068-8b21-e3afbde59e97-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:54 crc kubenswrapper[4702]: I1125 13:50:54.579751 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84464ec3-791f-4068-8b21-e3afbde59e97-config\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:54 crc kubenswrapper[4702]: I1125 13:50:54.579763 4702 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/84464ec3-791f-4068-8b21-e3afbde59e97-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.127930 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c5df878b4-4kk5g" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.127916 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c5df878b4-4kk5g" event={"ID":"97421c1d-3038-4993-8a2b-71b7629384fe","Type":"ContainerDied","Data":"26a433607349f18a8a4a6a9093011d8ef06d1807b1d01e52308c30ae075058a4"} Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.128396 4702 scope.go:117] "RemoveContainer" containerID="43265382df7d88cf229f78523838d46834521f8a7fddee8dc8115da420145574" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.129631 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-88759c5b4-xjptl" event={"ID":"84464ec3-791f-4068-8b21-e3afbde59e97","Type":"ContainerDied","Data":"78d05266ecf34af9f35184f384e62552a0ee898a906b1a69e5741a208c45f3b8"} Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.130289 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-88759c5b4-xjptl" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.144661 4702 scope.go:117] "RemoveContainer" containerID="94e7995139f29df1142a77015b54ddad293217bb450c398c99a266a8947f35a6" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.154930 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c5df878b4-4kk5g"] Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.159468 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c5df878b4-4kk5g"] Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.165340 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-88759c5b4-xjptl"] Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.169012 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-88759c5b4-xjptl"] Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.702228 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-95b665d9-2w2fh"] Nov 25 13:50:55 crc kubenswrapper[4702]: E1125 13:50:55.702478 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97421c1d-3038-4993-8a2b-71b7629384fe" containerName="route-controller-manager" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.702495 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="97421c1d-3038-4993-8a2b-71b7629384fe" containerName="route-controller-manager" Nov 25 13:50:55 crc kubenswrapper[4702]: E1125 13:50:55.702514 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84464ec3-791f-4068-8b21-e3afbde59e97" containerName="controller-manager" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.702522 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="84464ec3-791f-4068-8b21-e3afbde59e97" containerName="controller-manager" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.702619 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="84464ec3-791f-4068-8b21-e3afbde59e97" containerName="controller-manager" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.702639 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="97421c1d-3038-4993-8a2b-71b7629384fe" containerName="route-controller-manager" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.703088 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-95b665d9-2w2fh" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.707864 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.707942 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.707962 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.708590 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.709164 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.709261 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.726368 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6ccdd5cd9-8swj9"] Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.727205 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6ccdd5cd9-8swj9" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.729403 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.734933 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.735367 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.735467 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.735601 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.735698 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.735893 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.742214 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-95b665d9-2w2fh"] Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.762059 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6ccdd5cd9-8swj9"] Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.804962 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dh9k\" (UniqueName: \"kubernetes.io/projected/114a19ac-ef3a-4f27-8629-4faf56bb6943-kube-api-access-8dh9k\") pod \"controller-manager-95b665d9-2w2fh\" (UID: \"114a19ac-ef3a-4f27-8629-4faf56bb6943\") " pod="openshift-controller-manager/controller-manager-95b665d9-2w2fh" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.805042 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/114a19ac-ef3a-4f27-8629-4faf56bb6943-client-ca\") pod \"controller-manager-95b665d9-2w2fh\" (UID: \"114a19ac-ef3a-4f27-8629-4faf56bb6943\") " pod="openshift-controller-manager/controller-manager-95b665d9-2w2fh" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.805089 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/114a19ac-ef3a-4f27-8629-4faf56bb6943-serving-cert\") pod \"controller-manager-95b665d9-2w2fh\" (UID: \"114a19ac-ef3a-4f27-8629-4faf56bb6943\") " pod="openshift-controller-manager/controller-manager-95b665d9-2w2fh" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.805129 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/114a19ac-ef3a-4f27-8629-4faf56bb6943-config\") pod \"controller-manager-95b665d9-2w2fh\" (UID: \"114a19ac-ef3a-4f27-8629-4faf56bb6943\") " pod="openshift-controller-manager/controller-manager-95b665d9-2w2fh" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.805316 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/114a19ac-ef3a-4f27-8629-4faf56bb6943-proxy-ca-bundles\") pod \"controller-manager-95b665d9-2w2fh\" (UID: \"114a19ac-ef3a-4f27-8629-4faf56bb6943\") " pod="openshift-controller-manager/controller-manager-95b665d9-2w2fh" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.808691 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84464ec3-791f-4068-8b21-e3afbde59e97" path="/var/lib/kubelet/pods/84464ec3-791f-4068-8b21-e3afbde59e97/volumes" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.809384 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97421c1d-3038-4993-8a2b-71b7629384fe" path="/var/lib/kubelet/pods/97421c1d-3038-4993-8a2b-71b7629384fe/volumes" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.906326 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/114a19ac-ef3a-4f27-8629-4faf56bb6943-serving-cert\") pod \"controller-manager-95b665d9-2w2fh\" (UID: \"114a19ac-ef3a-4f27-8629-4faf56bb6943\") " pod="openshift-controller-manager/controller-manager-95b665d9-2w2fh" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.906386 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/114a19ac-ef3a-4f27-8629-4faf56bb6943-config\") pod \"controller-manager-95b665d9-2w2fh\" (UID: \"114a19ac-ef3a-4f27-8629-4faf56bb6943\") " pod="openshift-controller-manager/controller-manager-95b665d9-2w2fh" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.906415 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c-config\") pod \"route-controller-manager-6ccdd5cd9-8swj9\" (UID: \"e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c\") " pod="openshift-route-controller-manager/route-controller-manager-6ccdd5cd9-8swj9" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.906436 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bgz7\" (UniqueName: \"kubernetes.io/projected/e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c-kube-api-access-9bgz7\") pod \"route-controller-manager-6ccdd5cd9-8swj9\" (UID: \"e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c\") " pod="openshift-route-controller-manager/route-controller-manager-6ccdd5cd9-8swj9" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.906467 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/114a19ac-ef3a-4f27-8629-4faf56bb6943-proxy-ca-bundles\") pod \"controller-manager-95b665d9-2w2fh\" (UID: \"114a19ac-ef3a-4f27-8629-4faf56bb6943\") " pod="openshift-controller-manager/controller-manager-95b665d9-2w2fh" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.906484 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c-client-ca\") pod \"route-controller-manager-6ccdd5cd9-8swj9\" (UID: \"e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c\") " pod="openshift-route-controller-manager/route-controller-manager-6ccdd5cd9-8swj9" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.906540 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dh9k\" (UniqueName: \"kubernetes.io/projected/114a19ac-ef3a-4f27-8629-4faf56bb6943-kube-api-access-8dh9k\") pod \"controller-manager-95b665d9-2w2fh\" (UID: \"114a19ac-ef3a-4f27-8629-4faf56bb6943\") " pod="openshift-controller-manager/controller-manager-95b665d9-2w2fh" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.906556 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c-serving-cert\") pod \"route-controller-manager-6ccdd5cd9-8swj9\" (UID: \"e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c\") " pod="openshift-route-controller-manager/route-controller-manager-6ccdd5cd9-8swj9" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.906577 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/114a19ac-ef3a-4f27-8629-4faf56bb6943-client-ca\") pod \"controller-manager-95b665d9-2w2fh\" (UID: \"114a19ac-ef3a-4f27-8629-4faf56bb6943\") " pod="openshift-controller-manager/controller-manager-95b665d9-2w2fh" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.908018 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/114a19ac-ef3a-4f27-8629-4faf56bb6943-config\") pod \"controller-manager-95b665d9-2w2fh\" (UID: \"114a19ac-ef3a-4f27-8629-4faf56bb6943\") " pod="openshift-controller-manager/controller-manager-95b665d9-2w2fh" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.909458 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/114a19ac-ef3a-4f27-8629-4faf56bb6943-client-ca\") pod \"controller-manager-95b665d9-2w2fh\" (UID: \"114a19ac-ef3a-4f27-8629-4faf56bb6943\") " pod="openshift-controller-manager/controller-manager-95b665d9-2w2fh" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.909838 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/114a19ac-ef3a-4f27-8629-4faf56bb6943-proxy-ca-bundles\") pod \"controller-manager-95b665d9-2w2fh\" (UID: \"114a19ac-ef3a-4f27-8629-4faf56bb6943\") " pod="openshift-controller-manager/controller-manager-95b665d9-2w2fh" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.916018 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/114a19ac-ef3a-4f27-8629-4faf56bb6943-serving-cert\") pod \"controller-manager-95b665d9-2w2fh\" (UID: \"114a19ac-ef3a-4f27-8629-4faf56bb6943\") " pod="openshift-controller-manager/controller-manager-95b665d9-2w2fh" Nov 25 13:50:55 crc kubenswrapper[4702]: I1125 13:50:55.927137 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dh9k\" (UniqueName: \"kubernetes.io/projected/114a19ac-ef3a-4f27-8629-4faf56bb6943-kube-api-access-8dh9k\") pod \"controller-manager-95b665d9-2w2fh\" (UID: \"114a19ac-ef3a-4f27-8629-4faf56bb6943\") " pod="openshift-controller-manager/controller-manager-95b665d9-2w2fh" Nov 25 13:50:56 crc kubenswrapper[4702]: I1125 13:50:56.008310 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c-serving-cert\") pod \"route-controller-manager-6ccdd5cd9-8swj9\" (UID: \"e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c\") " pod="openshift-route-controller-manager/route-controller-manager-6ccdd5cd9-8swj9" Nov 25 13:50:56 crc kubenswrapper[4702]: I1125 13:50:56.008408 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c-config\") pod \"route-controller-manager-6ccdd5cd9-8swj9\" (UID: \"e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c\") " pod="openshift-route-controller-manager/route-controller-manager-6ccdd5cd9-8swj9" Nov 25 13:50:56 crc kubenswrapper[4702]: I1125 13:50:56.008433 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bgz7\" (UniqueName: \"kubernetes.io/projected/e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c-kube-api-access-9bgz7\") pod \"route-controller-manager-6ccdd5cd9-8swj9\" (UID: \"e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c\") " pod="openshift-route-controller-manager/route-controller-manager-6ccdd5cd9-8swj9" Nov 25 13:50:56 crc kubenswrapper[4702]: I1125 13:50:56.008457 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c-client-ca\") pod \"route-controller-manager-6ccdd5cd9-8swj9\" (UID: \"e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c\") " pod="openshift-route-controller-manager/route-controller-manager-6ccdd5cd9-8swj9" Nov 25 13:50:56 crc kubenswrapper[4702]: I1125 13:50:56.009159 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c-client-ca\") pod \"route-controller-manager-6ccdd5cd9-8swj9\" (UID: \"e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c\") " pod="openshift-route-controller-manager/route-controller-manager-6ccdd5cd9-8swj9" Nov 25 13:50:56 crc kubenswrapper[4702]: I1125 13:50:56.010488 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c-config\") pod \"route-controller-manager-6ccdd5cd9-8swj9\" (UID: \"e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c\") " pod="openshift-route-controller-manager/route-controller-manager-6ccdd5cd9-8swj9" Nov 25 13:50:56 crc kubenswrapper[4702]: I1125 13:50:56.012827 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c-serving-cert\") pod \"route-controller-manager-6ccdd5cd9-8swj9\" (UID: \"e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c\") " pod="openshift-route-controller-manager/route-controller-manager-6ccdd5cd9-8swj9" Nov 25 13:50:56 crc kubenswrapper[4702]: I1125 13:50:56.019033 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-95b665d9-2w2fh" Nov 25 13:50:56 crc kubenswrapper[4702]: I1125 13:50:56.038090 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bgz7\" (UniqueName: \"kubernetes.io/projected/e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c-kube-api-access-9bgz7\") pod \"route-controller-manager-6ccdd5cd9-8swj9\" (UID: \"e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c\") " pod="openshift-route-controller-manager/route-controller-manager-6ccdd5cd9-8swj9" Nov 25 13:50:56 crc kubenswrapper[4702]: I1125 13:50:56.043772 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6ccdd5cd9-8swj9" Nov 25 13:50:56 crc kubenswrapper[4702]: I1125 13:50:56.467551 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-95b665d9-2w2fh"] Nov 25 13:50:56 crc kubenswrapper[4702]: I1125 13:50:56.494684 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6ccdd5cd9-8swj9"] Nov 25 13:50:56 crc kubenswrapper[4702]: W1125 13:50:56.503023 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3cd004b_ae5e_4cfa_8496_1cf0fedacb8c.slice/crio-51ec9b1e566bb2abc6fc8a89d4238004d5107dd7c47243798558b4b87d4069f9 WatchSource:0}: Error finding container 51ec9b1e566bb2abc6fc8a89d4238004d5107dd7c47243798558b4b87d4069f9: Status 404 returned error can't find the container with id 51ec9b1e566bb2abc6fc8a89d4238004d5107dd7c47243798558b4b87d4069f9 Nov 25 13:50:57 crc kubenswrapper[4702]: I1125 13:50:57.164808 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-95b665d9-2w2fh" event={"ID":"114a19ac-ef3a-4f27-8629-4faf56bb6943","Type":"ContainerStarted","Data":"e994b359a7914e9ff3d40252fb13cad8312e29797145c0b3d8b72e4ae3b79c6f"} Nov 25 13:50:57 crc kubenswrapper[4702]: I1125 13:50:57.165789 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-95b665d9-2w2fh" event={"ID":"114a19ac-ef3a-4f27-8629-4faf56bb6943","Type":"ContainerStarted","Data":"ac1d7b6013969d14422d14bfcddb2506260c32c3bfd8b8a76d8bb45bcb7d8562"} Nov 25 13:50:57 crc kubenswrapper[4702]: I1125 13:50:57.166608 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-95b665d9-2w2fh" Nov 25 13:50:57 crc kubenswrapper[4702]: I1125 13:50:57.168822 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6ccdd5cd9-8swj9" event={"ID":"e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c","Type":"ContainerStarted","Data":"202057cb377ce8b78080795c951c2bd14cae0fdde47395222c4f5227f4cc8174"} Nov 25 13:50:57 crc kubenswrapper[4702]: I1125 13:50:57.168876 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6ccdd5cd9-8swj9" event={"ID":"e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c","Type":"ContainerStarted","Data":"51ec9b1e566bb2abc6fc8a89d4238004d5107dd7c47243798558b4b87d4069f9"} Nov 25 13:50:57 crc kubenswrapper[4702]: I1125 13:50:57.169902 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6ccdd5cd9-8swj9" Nov 25 13:50:57 crc kubenswrapper[4702]: I1125 13:50:57.174311 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-95b665d9-2w2fh" Nov 25 13:50:57 crc kubenswrapper[4702]: I1125 13:50:57.179048 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6ccdd5cd9-8swj9" Nov 25 13:50:57 crc kubenswrapper[4702]: I1125 13:50:57.196635 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-95b665d9-2w2fh" podStartSLOduration=2.196608095 podStartE2EDuration="2.196608095s" podCreationTimestamp="2025-11-25 13:50:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:50:57.190362266 +0000 UTC m=+266.353012716" watchObservedRunningTime="2025-11-25 13:50:57.196608095 +0000 UTC m=+266.359258545" Nov 25 13:50:57 crc kubenswrapper[4702]: I1125 13:50:57.215477 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6ccdd5cd9-8swj9" podStartSLOduration=2.215444192 podStartE2EDuration="2.215444192s" podCreationTimestamp="2025-11-25 13:50:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:50:57.212660637 +0000 UTC m=+266.375311097" watchObservedRunningTime="2025-11-25 13:50:57.215444192 +0000 UTC m=+266.378094642" Nov 25 13:51:18 crc kubenswrapper[4702]: I1125 13:51:18.127210 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5fmgv"] Nov 25 13:51:18 crc kubenswrapper[4702]: I1125 13:51:18.129097 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5fmgv" Nov 25 13:51:18 crc kubenswrapper[4702]: I1125 13:51:18.131622 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 13:51:18 crc kubenswrapper[4702]: I1125 13:51:18.138584 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5fmgv"] Nov 25 13:51:18 crc kubenswrapper[4702]: I1125 13:51:18.212894 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvcpc\" (UniqueName: \"kubernetes.io/projected/afc764af-1603-432c-9993-d0fc66fbb4fa-kube-api-access-nvcpc\") pod \"redhat-operators-5fmgv\" (UID: \"afc764af-1603-432c-9993-d0fc66fbb4fa\") " pod="openshift-marketplace/redhat-operators-5fmgv" Nov 25 13:51:18 crc kubenswrapper[4702]: I1125 13:51:18.212945 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afc764af-1603-432c-9993-d0fc66fbb4fa-utilities\") pod \"redhat-operators-5fmgv\" (UID: \"afc764af-1603-432c-9993-d0fc66fbb4fa\") " pod="openshift-marketplace/redhat-operators-5fmgv" Nov 25 13:51:18 crc kubenswrapper[4702]: I1125 13:51:18.212988 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afc764af-1603-432c-9993-d0fc66fbb4fa-catalog-content\") pod \"redhat-operators-5fmgv\" (UID: \"afc764af-1603-432c-9993-d0fc66fbb4fa\") " pod="openshift-marketplace/redhat-operators-5fmgv" Nov 25 13:51:18 crc kubenswrapper[4702]: I1125 13:51:18.314013 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvcpc\" (UniqueName: \"kubernetes.io/projected/afc764af-1603-432c-9993-d0fc66fbb4fa-kube-api-access-nvcpc\") pod \"redhat-operators-5fmgv\" (UID: \"afc764af-1603-432c-9993-d0fc66fbb4fa\") " pod="openshift-marketplace/redhat-operators-5fmgv" Nov 25 13:51:18 crc kubenswrapper[4702]: I1125 13:51:18.314071 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afc764af-1603-432c-9993-d0fc66fbb4fa-utilities\") pod \"redhat-operators-5fmgv\" (UID: \"afc764af-1603-432c-9993-d0fc66fbb4fa\") " pod="openshift-marketplace/redhat-operators-5fmgv" Nov 25 13:51:18 crc kubenswrapper[4702]: I1125 13:51:18.314124 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afc764af-1603-432c-9993-d0fc66fbb4fa-catalog-content\") pod \"redhat-operators-5fmgv\" (UID: \"afc764af-1603-432c-9993-d0fc66fbb4fa\") " pod="openshift-marketplace/redhat-operators-5fmgv" Nov 25 13:51:18 crc kubenswrapper[4702]: I1125 13:51:18.314671 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afc764af-1603-432c-9993-d0fc66fbb4fa-utilities\") pod \"redhat-operators-5fmgv\" (UID: \"afc764af-1603-432c-9993-d0fc66fbb4fa\") " pod="openshift-marketplace/redhat-operators-5fmgv" Nov 25 13:51:18 crc kubenswrapper[4702]: I1125 13:51:18.314714 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afc764af-1603-432c-9993-d0fc66fbb4fa-catalog-content\") pod \"redhat-operators-5fmgv\" (UID: \"afc764af-1603-432c-9993-d0fc66fbb4fa\") " pod="openshift-marketplace/redhat-operators-5fmgv" Nov 25 13:51:18 crc kubenswrapper[4702]: I1125 13:51:18.334986 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvcpc\" (UniqueName: \"kubernetes.io/projected/afc764af-1603-432c-9993-d0fc66fbb4fa-kube-api-access-nvcpc\") pod \"redhat-operators-5fmgv\" (UID: \"afc764af-1603-432c-9993-d0fc66fbb4fa\") " pod="openshift-marketplace/redhat-operators-5fmgv" Nov 25 13:51:18 crc kubenswrapper[4702]: I1125 13:51:18.448605 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5fmgv" Nov 25 13:51:18 crc kubenswrapper[4702]: I1125 13:51:18.724656 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-r28xx"] Nov 25 13:51:18 crc kubenswrapper[4702]: I1125 13:51:18.726222 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r28xx" Nov 25 13:51:18 crc kubenswrapper[4702]: I1125 13:51:18.728498 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 25 13:51:18 crc kubenswrapper[4702]: I1125 13:51:18.736586 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r28xx"] Nov 25 13:51:18 crc kubenswrapper[4702]: I1125 13:51:18.824204 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fj479\" (UniqueName: \"kubernetes.io/projected/2b30ed31-5291-4679-89a2-ca54d9f477b8-kube-api-access-fj479\") pod \"community-operators-r28xx\" (UID: \"2b30ed31-5291-4679-89a2-ca54d9f477b8\") " pod="openshift-marketplace/community-operators-r28xx" Nov 25 13:51:18 crc kubenswrapper[4702]: I1125 13:51:18.824437 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b30ed31-5291-4679-89a2-ca54d9f477b8-catalog-content\") pod \"community-operators-r28xx\" (UID: \"2b30ed31-5291-4679-89a2-ca54d9f477b8\") " pod="openshift-marketplace/community-operators-r28xx" Nov 25 13:51:18 crc kubenswrapper[4702]: I1125 13:51:18.824494 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b30ed31-5291-4679-89a2-ca54d9f477b8-utilities\") pod \"community-operators-r28xx\" (UID: \"2b30ed31-5291-4679-89a2-ca54d9f477b8\") " pod="openshift-marketplace/community-operators-r28xx" Nov 25 13:51:18 crc kubenswrapper[4702]: I1125 13:51:18.837381 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5fmgv"] Nov 25 13:51:18 crc kubenswrapper[4702]: W1125 13:51:18.840904 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podafc764af_1603_432c_9993_d0fc66fbb4fa.slice/crio-95c6a1004660aee7a569c9e13360d6ce6acbe2bb3da1ed605056793d0adc0a76 WatchSource:0}: Error finding container 95c6a1004660aee7a569c9e13360d6ce6acbe2bb3da1ed605056793d0adc0a76: Status 404 returned error can't find the container with id 95c6a1004660aee7a569c9e13360d6ce6acbe2bb3da1ed605056793d0adc0a76 Nov 25 13:51:18 crc kubenswrapper[4702]: I1125 13:51:18.925098 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b30ed31-5291-4679-89a2-ca54d9f477b8-catalog-content\") pod \"community-operators-r28xx\" (UID: \"2b30ed31-5291-4679-89a2-ca54d9f477b8\") " pod="openshift-marketplace/community-operators-r28xx" Nov 25 13:51:18 crc kubenswrapper[4702]: I1125 13:51:18.925147 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b30ed31-5291-4679-89a2-ca54d9f477b8-utilities\") pod \"community-operators-r28xx\" (UID: \"2b30ed31-5291-4679-89a2-ca54d9f477b8\") " pod="openshift-marketplace/community-operators-r28xx" Nov 25 13:51:18 crc kubenswrapper[4702]: I1125 13:51:18.925179 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fj479\" (UniqueName: \"kubernetes.io/projected/2b30ed31-5291-4679-89a2-ca54d9f477b8-kube-api-access-fj479\") pod \"community-operators-r28xx\" (UID: \"2b30ed31-5291-4679-89a2-ca54d9f477b8\") " pod="openshift-marketplace/community-operators-r28xx" Nov 25 13:51:18 crc kubenswrapper[4702]: I1125 13:51:18.926043 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b30ed31-5291-4679-89a2-ca54d9f477b8-catalog-content\") pod \"community-operators-r28xx\" (UID: \"2b30ed31-5291-4679-89a2-ca54d9f477b8\") " pod="openshift-marketplace/community-operators-r28xx" Nov 25 13:51:18 crc kubenswrapper[4702]: I1125 13:51:18.926116 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b30ed31-5291-4679-89a2-ca54d9f477b8-utilities\") pod \"community-operators-r28xx\" (UID: \"2b30ed31-5291-4679-89a2-ca54d9f477b8\") " pod="openshift-marketplace/community-operators-r28xx" Nov 25 13:51:18 crc kubenswrapper[4702]: I1125 13:51:18.945282 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fj479\" (UniqueName: \"kubernetes.io/projected/2b30ed31-5291-4679-89a2-ca54d9f477b8-kube-api-access-fj479\") pod \"community-operators-r28xx\" (UID: \"2b30ed31-5291-4679-89a2-ca54d9f477b8\") " pod="openshift-marketplace/community-operators-r28xx" Nov 25 13:51:18 crc kubenswrapper[4702]: I1125 13:51:18.967414 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-k8g4d"] Nov 25 13:51:18 crc kubenswrapper[4702]: I1125 13:51:18.968105 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-k8g4d" Nov 25 13:51:19 crc kubenswrapper[4702]: I1125 13:51:18.982107 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-k8g4d"] Nov 25 13:51:19 crc kubenswrapper[4702]: I1125 13:51:19.049534 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r28xx" Nov 25 13:51:19 crc kubenswrapper[4702]: I1125 13:51:19.128897 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/45eeab18-b29a-4ffd-9120-dafa607d14f3-registry-tls\") pod \"image-registry-66df7c8f76-k8g4d\" (UID: \"45eeab18-b29a-4ffd-9120-dafa607d14f3\") " pod="openshift-image-registry/image-registry-66df7c8f76-k8g4d" Nov 25 13:51:19 crc kubenswrapper[4702]: I1125 13:51:19.128943 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/45eeab18-b29a-4ffd-9120-dafa607d14f3-installation-pull-secrets\") pod \"image-registry-66df7c8f76-k8g4d\" (UID: \"45eeab18-b29a-4ffd-9120-dafa607d14f3\") " pod="openshift-image-registry/image-registry-66df7c8f76-k8g4d" Nov 25 13:51:19 crc kubenswrapper[4702]: I1125 13:51:19.128992 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-k8g4d\" (UID: \"45eeab18-b29a-4ffd-9120-dafa607d14f3\") " pod="openshift-image-registry/image-registry-66df7c8f76-k8g4d" Nov 25 13:51:19 crc kubenswrapper[4702]: I1125 13:51:19.129017 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/45eeab18-b29a-4ffd-9120-dafa607d14f3-trusted-ca\") pod \"image-registry-66df7c8f76-k8g4d\" (UID: \"45eeab18-b29a-4ffd-9120-dafa607d14f3\") " pod="openshift-image-registry/image-registry-66df7c8f76-k8g4d" Nov 25 13:51:19 crc kubenswrapper[4702]: I1125 13:51:19.129057 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6q7lj\" (UniqueName: \"kubernetes.io/projected/45eeab18-b29a-4ffd-9120-dafa607d14f3-kube-api-access-6q7lj\") pod \"image-registry-66df7c8f76-k8g4d\" (UID: \"45eeab18-b29a-4ffd-9120-dafa607d14f3\") " pod="openshift-image-registry/image-registry-66df7c8f76-k8g4d" Nov 25 13:51:19 crc kubenswrapper[4702]: I1125 13:51:19.129077 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/45eeab18-b29a-4ffd-9120-dafa607d14f3-registry-certificates\") pod \"image-registry-66df7c8f76-k8g4d\" (UID: \"45eeab18-b29a-4ffd-9120-dafa607d14f3\") " pod="openshift-image-registry/image-registry-66df7c8f76-k8g4d" Nov 25 13:51:19 crc kubenswrapper[4702]: I1125 13:51:19.129105 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/45eeab18-b29a-4ffd-9120-dafa607d14f3-bound-sa-token\") pod \"image-registry-66df7c8f76-k8g4d\" (UID: \"45eeab18-b29a-4ffd-9120-dafa607d14f3\") " pod="openshift-image-registry/image-registry-66df7c8f76-k8g4d" Nov 25 13:51:19 crc kubenswrapper[4702]: I1125 13:51:19.129128 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/45eeab18-b29a-4ffd-9120-dafa607d14f3-ca-trust-extracted\") pod \"image-registry-66df7c8f76-k8g4d\" (UID: \"45eeab18-b29a-4ffd-9120-dafa607d14f3\") " pod="openshift-image-registry/image-registry-66df7c8f76-k8g4d" Nov 25 13:51:19 crc kubenswrapper[4702]: I1125 13:51:19.159314 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-k8g4d\" (UID: \"45eeab18-b29a-4ffd-9120-dafa607d14f3\") " pod="openshift-image-registry/image-registry-66df7c8f76-k8g4d" Nov 25 13:51:19 crc kubenswrapper[4702]: I1125 13:51:19.230342 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/45eeab18-b29a-4ffd-9120-dafa607d14f3-registry-tls\") pod \"image-registry-66df7c8f76-k8g4d\" (UID: \"45eeab18-b29a-4ffd-9120-dafa607d14f3\") " pod="openshift-image-registry/image-registry-66df7c8f76-k8g4d" Nov 25 13:51:19 crc kubenswrapper[4702]: I1125 13:51:19.230384 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/45eeab18-b29a-4ffd-9120-dafa607d14f3-installation-pull-secrets\") pod \"image-registry-66df7c8f76-k8g4d\" (UID: \"45eeab18-b29a-4ffd-9120-dafa607d14f3\") " pod="openshift-image-registry/image-registry-66df7c8f76-k8g4d" Nov 25 13:51:19 crc kubenswrapper[4702]: I1125 13:51:19.230442 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/45eeab18-b29a-4ffd-9120-dafa607d14f3-trusted-ca\") pod \"image-registry-66df7c8f76-k8g4d\" (UID: \"45eeab18-b29a-4ffd-9120-dafa607d14f3\") " pod="openshift-image-registry/image-registry-66df7c8f76-k8g4d" Nov 25 13:51:19 crc kubenswrapper[4702]: I1125 13:51:19.230499 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6q7lj\" (UniqueName: \"kubernetes.io/projected/45eeab18-b29a-4ffd-9120-dafa607d14f3-kube-api-access-6q7lj\") pod \"image-registry-66df7c8f76-k8g4d\" (UID: \"45eeab18-b29a-4ffd-9120-dafa607d14f3\") " pod="openshift-image-registry/image-registry-66df7c8f76-k8g4d" Nov 25 13:51:19 crc kubenswrapper[4702]: I1125 13:51:19.230518 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/45eeab18-b29a-4ffd-9120-dafa607d14f3-registry-certificates\") pod \"image-registry-66df7c8f76-k8g4d\" (UID: \"45eeab18-b29a-4ffd-9120-dafa607d14f3\") " pod="openshift-image-registry/image-registry-66df7c8f76-k8g4d" Nov 25 13:51:19 crc kubenswrapper[4702]: I1125 13:51:19.230540 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/45eeab18-b29a-4ffd-9120-dafa607d14f3-bound-sa-token\") pod \"image-registry-66df7c8f76-k8g4d\" (UID: \"45eeab18-b29a-4ffd-9120-dafa607d14f3\") " pod="openshift-image-registry/image-registry-66df7c8f76-k8g4d" Nov 25 13:51:19 crc kubenswrapper[4702]: I1125 13:51:19.230559 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/45eeab18-b29a-4ffd-9120-dafa607d14f3-ca-trust-extracted\") pod \"image-registry-66df7c8f76-k8g4d\" (UID: \"45eeab18-b29a-4ffd-9120-dafa607d14f3\") " pod="openshift-image-registry/image-registry-66df7c8f76-k8g4d" Nov 25 13:51:19 crc kubenswrapper[4702]: I1125 13:51:19.232173 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/45eeab18-b29a-4ffd-9120-dafa607d14f3-trusted-ca\") pod \"image-registry-66df7c8f76-k8g4d\" (UID: \"45eeab18-b29a-4ffd-9120-dafa607d14f3\") " pod="openshift-image-registry/image-registry-66df7c8f76-k8g4d" Nov 25 13:51:19 crc kubenswrapper[4702]: I1125 13:51:19.232682 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/45eeab18-b29a-4ffd-9120-dafa607d14f3-registry-certificates\") pod \"image-registry-66df7c8f76-k8g4d\" (UID: \"45eeab18-b29a-4ffd-9120-dafa607d14f3\") " pod="openshift-image-registry/image-registry-66df7c8f76-k8g4d" Nov 25 13:51:19 crc kubenswrapper[4702]: I1125 13:51:19.233429 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/45eeab18-b29a-4ffd-9120-dafa607d14f3-ca-trust-extracted\") pod \"image-registry-66df7c8f76-k8g4d\" (UID: \"45eeab18-b29a-4ffd-9120-dafa607d14f3\") " pod="openshift-image-registry/image-registry-66df7c8f76-k8g4d" Nov 25 13:51:19 crc kubenswrapper[4702]: I1125 13:51:19.235604 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/45eeab18-b29a-4ffd-9120-dafa607d14f3-installation-pull-secrets\") pod \"image-registry-66df7c8f76-k8g4d\" (UID: \"45eeab18-b29a-4ffd-9120-dafa607d14f3\") " pod="openshift-image-registry/image-registry-66df7c8f76-k8g4d" Nov 25 13:51:19 crc kubenswrapper[4702]: I1125 13:51:19.247306 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/45eeab18-b29a-4ffd-9120-dafa607d14f3-registry-tls\") pod \"image-registry-66df7c8f76-k8g4d\" (UID: \"45eeab18-b29a-4ffd-9120-dafa607d14f3\") " pod="openshift-image-registry/image-registry-66df7c8f76-k8g4d" Nov 25 13:51:19 crc kubenswrapper[4702]: I1125 13:51:19.249442 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/45eeab18-b29a-4ffd-9120-dafa607d14f3-bound-sa-token\") pod \"image-registry-66df7c8f76-k8g4d\" (UID: \"45eeab18-b29a-4ffd-9120-dafa607d14f3\") " pod="openshift-image-registry/image-registry-66df7c8f76-k8g4d" Nov 25 13:51:19 crc kubenswrapper[4702]: I1125 13:51:19.252070 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6q7lj\" (UniqueName: \"kubernetes.io/projected/45eeab18-b29a-4ffd-9120-dafa607d14f3-kube-api-access-6q7lj\") pod \"image-registry-66df7c8f76-k8g4d\" (UID: \"45eeab18-b29a-4ffd-9120-dafa607d14f3\") " pod="openshift-image-registry/image-registry-66df7c8f76-k8g4d" Nov 25 13:51:19 crc kubenswrapper[4702]: I1125 13:51:19.281152 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-k8g4d" Nov 25 13:51:19 crc kubenswrapper[4702]: I1125 13:51:19.286557 4702 generic.go:334] "Generic (PLEG): container finished" podID="afc764af-1603-432c-9993-d0fc66fbb4fa" containerID="b07c4c61e5b3c919abbfbb7fe6fca7436f938cf26963d4f2ff7e5aee8bd10925" exitCode=0 Nov 25 13:51:19 crc kubenswrapper[4702]: I1125 13:51:19.286592 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5fmgv" event={"ID":"afc764af-1603-432c-9993-d0fc66fbb4fa","Type":"ContainerDied","Data":"b07c4c61e5b3c919abbfbb7fe6fca7436f938cf26963d4f2ff7e5aee8bd10925"} Nov 25 13:51:19 crc kubenswrapper[4702]: I1125 13:51:19.286615 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5fmgv" event={"ID":"afc764af-1603-432c-9993-d0fc66fbb4fa","Type":"ContainerStarted","Data":"95c6a1004660aee7a569c9e13360d6ce6acbe2bb3da1ed605056793d0adc0a76"} Nov 25 13:51:20 crc kubenswrapper[4702]: E1125 13:51:20.882012 4702 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.082s" Nov 25 13:51:20 crc kubenswrapper[4702]: I1125 13:51:20.882077 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xlvf4"] Nov 25 13:51:20 crc kubenswrapper[4702]: I1125 13:51:20.883394 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xlvf4"] Nov 25 13:51:20 crc kubenswrapper[4702]: I1125 13:51:20.883498 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xlvf4" Nov 25 13:51:20 crc kubenswrapper[4702]: I1125 13:51:20.886301 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 13:51:20 crc kubenswrapper[4702]: I1125 13:51:20.888385 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r28xx"] Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.059342 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/000d8b95-f22b-490b-b418-b8a070ed69f8-catalog-content\") pod \"certified-operators-xlvf4\" (UID: \"000d8b95-f22b-490b-b418-b8a070ed69f8\") " pod="openshift-marketplace/certified-operators-xlvf4" Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.059944 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/000d8b95-f22b-490b-b418-b8a070ed69f8-utilities\") pod \"certified-operators-xlvf4\" (UID: \"000d8b95-f22b-490b-b418-b8a070ed69f8\") " pod="openshift-marketplace/certified-operators-xlvf4" Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.060004 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwlsk\" (UniqueName: \"kubernetes.io/projected/000d8b95-f22b-490b-b418-b8a070ed69f8-kube-api-access-nwlsk\") pod \"certified-operators-xlvf4\" (UID: \"000d8b95-f22b-490b-b418-b8a070ed69f8\") " pod="openshift-marketplace/certified-operators-xlvf4" Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.121352 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nhp97"] Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.122539 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nhp97" Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.124812 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.127707 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nhp97"] Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.173907 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/000d8b95-f22b-490b-b418-b8a070ed69f8-utilities\") pod \"certified-operators-xlvf4\" (UID: \"000d8b95-f22b-490b-b418-b8a070ed69f8\") " pod="openshift-marketplace/certified-operators-xlvf4" Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.174003 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zp7vm\" (UniqueName: \"kubernetes.io/projected/ecf07980-291f-474e-9a27-9f050288a987-kube-api-access-zp7vm\") pod \"redhat-marketplace-nhp97\" (UID: \"ecf07980-291f-474e-9a27-9f050288a987\") " pod="openshift-marketplace/redhat-marketplace-nhp97" Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.174035 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwlsk\" (UniqueName: \"kubernetes.io/projected/000d8b95-f22b-490b-b418-b8a070ed69f8-kube-api-access-nwlsk\") pod \"certified-operators-xlvf4\" (UID: \"000d8b95-f22b-490b-b418-b8a070ed69f8\") " pod="openshift-marketplace/certified-operators-xlvf4" Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.174058 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecf07980-291f-474e-9a27-9f050288a987-utilities\") pod \"redhat-marketplace-nhp97\" (UID: \"ecf07980-291f-474e-9a27-9f050288a987\") " pod="openshift-marketplace/redhat-marketplace-nhp97" Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.174095 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecf07980-291f-474e-9a27-9f050288a987-catalog-content\") pod \"redhat-marketplace-nhp97\" (UID: \"ecf07980-291f-474e-9a27-9f050288a987\") " pod="openshift-marketplace/redhat-marketplace-nhp97" Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.174116 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/000d8b95-f22b-490b-b418-b8a070ed69f8-catalog-content\") pod \"certified-operators-xlvf4\" (UID: \"000d8b95-f22b-490b-b418-b8a070ed69f8\") " pod="openshift-marketplace/certified-operators-xlvf4" Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.174625 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/000d8b95-f22b-490b-b418-b8a070ed69f8-catalog-content\") pod \"certified-operators-xlvf4\" (UID: \"000d8b95-f22b-490b-b418-b8a070ed69f8\") " pod="openshift-marketplace/certified-operators-xlvf4" Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.174724 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/000d8b95-f22b-490b-b418-b8a070ed69f8-utilities\") pod \"certified-operators-xlvf4\" (UID: \"000d8b95-f22b-490b-b418-b8a070ed69f8\") " pod="openshift-marketplace/certified-operators-xlvf4" Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.201505 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwlsk\" (UniqueName: \"kubernetes.io/projected/000d8b95-f22b-490b-b418-b8a070ed69f8-kube-api-access-nwlsk\") pod \"certified-operators-xlvf4\" (UID: \"000d8b95-f22b-490b-b418-b8a070ed69f8\") " pod="openshift-marketplace/certified-operators-xlvf4" Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.235756 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xlvf4" Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.235963 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-k8g4d"] Nov 25 13:51:21 crc kubenswrapper[4702]: W1125 13:51:21.240734 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45eeab18_b29a_4ffd_9120_dafa607d14f3.slice/crio-e7f88cf2177628e1c8a01e24dea00ee307f098035afd0933b663aa1fc77479aa WatchSource:0}: Error finding container e7f88cf2177628e1c8a01e24dea00ee307f098035afd0933b663aa1fc77479aa: Status 404 returned error can't find the container with id e7f88cf2177628e1c8a01e24dea00ee307f098035afd0933b663aa1fc77479aa Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.275041 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zp7vm\" (UniqueName: \"kubernetes.io/projected/ecf07980-291f-474e-9a27-9f050288a987-kube-api-access-zp7vm\") pod \"redhat-marketplace-nhp97\" (UID: \"ecf07980-291f-474e-9a27-9f050288a987\") " pod="openshift-marketplace/redhat-marketplace-nhp97" Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.275125 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecf07980-291f-474e-9a27-9f050288a987-utilities\") pod \"redhat-marketplace-nhp97\" (UID: \"ecf07980-291f-474e-9a27-9f050288a987\") " pod="openshift-marketplace/redhat-marketplace-nhp97" Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.275179 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecf07980-291f-474e-9a27-9f050288a987-catalog-content\") pod \"redhat-marketplace-nhp97\" (UID: \"ecf07980-291f-474e-9a27-9f050288a987\") " pod="openshift-marketplace/redhat-marketplace-nhp97" Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.275962 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecf07980-291f-474e-9a27-9f050288a987-catalog-content\") pod \"redhat-marketplace-nhp97\" (UID: \"ecf07980-291f-474e-9a27-9f050288a987\") " pod="openshift-marketplace/redhat-marketplace-nhp97" Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.277857 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecf07980-291f-474e-9a27-9f050288a987-utilities\") pod \"redhat-marketplace-nhp97\" (UID: \"ecf07980-291f-474e-9a27-9f050288a987\") " pod="openshift-marketplace/redhat-marketplace-nhp97" Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.296171 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zp7vm\" (UniqueName: \"kubernetes.io/projected/ecf07980-291f-474e-9a27-9f050288a987-kube-api-access-zp7vm\") pod \"redhat-marketplace-nhp97\" (UID: \"ecf07980-291f-474e-9a27-9f050288a987\") " pod="openshift-marketplace/redhat-marketplace-nhp97" Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.528096 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nhp97" Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.635607 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xlvf4"] Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.891007 4702 generic.go:334] "Generic (PLEG): container finished" podID="000d8b95-f22b-490b-b418-b8a070ed69f8" containerID="8cd36b95390074930aaaa247753e6216075b4cbd41dc10a57f4af560ce117134" exitCode=0 Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.891554 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xlvf4" event={"ID":"000d8b95-f22b-490b-b418-b8a070ed69f8","Type":"ContainerDied","Data":"8cd36b95390074930aaaa247753e6216075b4cbd41dc10a57f4af560ce117134"} Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.892067 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xlvf4" event={"ID":"000d8b95-f22b-490b-b418-b8a070ed69f8","Type":"ContainerStarted","Data":"75a955f034110834c18261dc9dbaf60196ed936da599a73d82243febe58c2ae8"} Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.893616 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-k8g4d" event={"ID":"45eeab18-b29a-4ffd-9120-dafa607d14f3","Type":"ContainerStarted","Data":"f42c72f7ed200e61ffeaf1243298e75696d373209ac50713edc08458365f49fc"} Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.893641 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-k8g4d" event={"ID":"45eeab18-b29a-4ffd-9120-dafa607d14f3","Type":"ContainerStarted","Data":"e7f88cf2177628e1c8a01e24dea00ee307f098035afd0933b663aa1fc77479aa"} Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.894423 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-k8g4d" Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.897226 4702 generic.go:334] "Generic (PLEG): container finished" podID="2b30ed31-5291-4679-89a2-ca54d9f477b8" containerID="f55e92112612baf0775861331e0b849dddea30c3c2a30c1f26e50f70a7366cd5" exitCode=0 Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.898109 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r28xx" event={"ID":"2b30ed31-5291-4679-89a2-ca54d9f477b8","Type":"ContainerDied","Data":"f55e92112612baf0775861331e0b849dddea30c3c2a30c1f26e50f70a7366cd5"} Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.898152 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r28xx" event={"ID":"2b30ed31-5291-4679-89a2-ca54d9f477b8","Type":"ContainerStarted","Data":"066f2c3705610d2f1117bab4b2586cebed4c960f4755aaa364c9d53ab0999d59"} Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.960570 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-k8g4d" podStartSLOduration=3.960545224 podStartE2EDuration="3.960545224s" podCreationTimestamp="2025-11-25 13:51:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:51:21.955043155 +0000 UTC m=+291.117693605" watchObservedRunningTime="2025-11-25 13:51:21.960545224 +0000 UTC m=+291.123195674" Nov 25 13:51:21 crc kubenswrapper[4702]: I1125 13:51:21.971711 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nhp97"] Nov 25 13:51:22 crc kubenswrapper[4702]: I1125 13:51:22.905158 4702 generic.go:334] "Generic (PLEG): container finished" podID="ecf07980-291f-474e-9a27-9f050288a987" containerID="d727cd6a6ac023a75f054369999225ceeb42454c4976834b4b02c4d9563e1630" exitCode=0 Nov 25 13:51:22 crc kubenswrapper[4702]: I1125 13:51:22.905261 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nhp97" event={"ID":"ecf07980-291f-474e-9a27-9f050288a987","Type":"ContainerDied","Data":"d727cd6a6ac023a75f054369999225ceeb42454c4976834b4b02c4d9563e1630"} Nov 25 13:51:22 crc kubenswrapper[4702]: I1125 13:51:22.906011 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nhp97" event={"ID":"ecf07980-291f-474e-9a27-9f050288a987","Type":"ContainerStarted","Data":"b6fd672c664faf2abfecffa6f732f0f9e166da3aeb826f2fda118a833c3616b0"} Nov 25 13:51:23 crc kubenswrapper[4702]: I1125 13:51:23.912468 4702 generic.go:334] "Generic (PLEG): container finished" podID="000d8b95-f22b-490b-b418-b8a070ed69f8" containerID="6149e76b5911e43e58ef0b09e9dd9d05368a3bbebf6124e75a7e052408a83d49" exitCode=0 Nov 25 13:51:23 crc kubenswrapper[4702]: I1125 13:51:23.912531 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xlvf4" event={"ID":"000d8b95-f22b-490b-b418-b8a070ed69f8","Type":"ContainerDied","Data":"6149e76b5911e43e58ef0b09e9dd9d05368a3bbebf6124e75a7e052408a83d49"} Nov 25 13:51:23 crc kubenswrapper[4702]: I1125 13:51:23.915846 4702 generic.go:334] "Generic (PLEG): container finished" podID="2b30ed31-5291-4679-89a2-ca54d9f477b8" containerID="f9ba3fe104a24300efe2cf4ef2265b677e41f0c13018cf25cfa76995d72abbc8" exitCode=0 Nov 25 13:51:23 crc kubenswrapper[4702]: I1125 13:51:23.915935 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r28xx" event={"ID":"2b30ed31-5291-4679-89a2-ca54d9f477b8","Type":"ContainerDied","Data":"f9ba3fe104a24300efe2cf4ef2265b677e41f0c13018cf25cfa76995d72abbc8"} Nov 25 13:51:23 crc kubenswrapper[4702]: I1125 13:51:23.919596 4702 generic.go:334] "Generic (PLEG): container finished" podID="afc764af-1603-432c-9993-d0fc66fbb4fa" containerID="58d9999d5cdc4f5abe0d62fe2f5633db3a8493f69eada9cc20959117910dd86d" exitCode=0 Nov 25 13:51:23 crc kubenswrapper[4702]: I1125 13:51:23.919640 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5fmgv" event={"ID":"afc764af-1603-432c-9993-d0fc66fbb4fa","Type":"ContainerDied","Data":"58d9999d5cdc4f5abe0d62fe2f5633db3a8493f69eada9cc20959117910dd86d"} Nov 25 13:51:23 crc kubenswrapper[4702]: I1125 13:51:23.921772 4702 generic.go:334] "Generic (PLEG): container finished" podID="ecf07980-291f-474e-9a27-9f050288a987" containerID="4212df1b876d3fea62125f201cf9e5f9bf93207f9eec79aba13e6618423123ce" exitCode=0 Nov 25 13:51:23 crc kubenswrapper[4702]: I1125 13:51:23.922315 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nhp97" event={"ID":"ecf07980-291f-474e-9a27-9f050288a987","Type":"ContainerDied","Data":"4212df1b876d3fea62125f201cf9e5f9bf93207f9eec79aba13e6618423123ce"} Nov 25 13:51:24 crc kubenswrapper[4702]: I1125 13:51:24.932529 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nhp97" event={"ID":"ecf07980-291f-474e-9a27-9f050288a987","Type":"ContainerStarted","Data":"4a96c4b0f908e923837405104aeeb1ed407961cc89e0326b2de4601a5e233e45"} Nov 25 13:51:24 crc kubenswrapper[4702]: I1125 13:51:24.936593 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xlvf4" event={"ID":"000d8b95-f22b-490b-b418-b8a070ed69f8","Type":"ContainerStarted","Data":"8ad2c179a560e90d7b97e54a9881cd29453ebb15c2920ab0b7088c156471444f"} Nov 25 13:51:24 crc kubenswrapper[4702]: I1125 13:51:24.939062 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r28xx" event={"ID":"2b30ed31-5291-4679-89a2-ca54d9f477b8","Type":"ContainerStarted","Data":"3ad2f21ac63b821b9aa0dc7f2b73f39a42388fc59e154115d1b8af8676e040d6"} Nov 25 13:51:24 crc kubenswrapper[4702]: I1125 13:51:24.941416 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5fmgv" event={"ID":"afc764af-1603-432c-9993-d0fc66fbb4fa","Type":"ContainerStarted","Data":"d61654ec1733f24d2c15e811d383de640608b8536f5ced3b3771cd04f59b18e1"} Nov 25 13:51:24 crc kubenswrapper[4702]: I1125 13:51:24.960022 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nhp97" podStartSLOduration=2.535748031 podStartE2EDuration="3.959999815s" podCreationTimestamp="2025-11-25 13:51:21 +0000 UTC" firstStartedPulling="2025-11-25 13:51:22.907849945 +0000 UTC m=+292.070500435" lastFinishedPulling="2025-11-25 13:51:24.332101769 +0000 UTC m=+293.494752219" observedRunningTime="2025-11-25 13:51:24.956964592 +0000 UTC m=+294.119615062" watchObservedRunningTime="2025-11-25 13:51:24.959999815 +0000 UTC m=+294.122650265" Nov 25 13:51:24 crc kubenswrapper[4702]: I1125 13:51:24.984591 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5fmgv" podStartSLOduration=4.417433166 podStartE2EDuration="6.984570757s" podCreationTimestamp="2025-11-25 13:51:18 +0000 UTC" firstStartedPulling="2025-11-25 13:51:21.898599933 +0000 UTC m=+291.061250383" lastFinishedPulling="2025-11-25 13:51:24.465737524 +0000 UTC m=+293.628387974" observedRunningTime="2025-11-25 13:51:24.978961856 +0000 UTC m=+294.141612306" watchObservedRunningTime="2025-11-25 13:51:24.984570757 +0000 UTC m=+294.147221207" Nov 25 13:51:24 crc kubenswrapper[4702]: I1125 13:51:24.997654 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-r28xx" podStartSLOduration=4.477485957 podStartE2EDuration="6.9976412s" podCreationTimestamp="2025-11-25 13:51:18 +0000 UTC" firstStartedPulling="2025-11-25 13:51:21.899290922 +0000 UTC m=+291.061941372" lastFinishedPulling="2025-11-25 13:51:24.419446165 +0000 UTC m=+293.582096615" observedRunningTime="2025-11-25 13:51:24.996813708 +0000 UTC m=+294.159464158" watchObservedRunningTime="2025-11-25 13:51:24.9976412 +0000 UTC m=+294.160291650" Nov 25 13:51:25 crc kubenswrapper[4702]: I1125 13:51:25.017629 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xlvf4" podStartSLOduration=2.429778821 podStartE2EDuration="5.017608009s" podCreationTimestamp="2025-11-25 13:51:20 +0000 UTC" firstStartedPulling="2025-11-25 13:51:21.894551134 +0000 UTC m=+291.057201584" lastFinishedPulling="2025-11-25 13:51:24.482380312 +0000 UTC m=+293.645030772" observedRunningTime="2025-11-25 13:51:25.01618412 +0000 UTC m=+294.178834570" watchObservedRunningTime="2025-11-25 13:51:25.017608009 +0000 UTC m=+294.180258459" Nov 25 13:51:27 crc kubenswrapper[4702]: I1125 13:51:27.534097 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-95b665d9-2w2fh"] Nov 25 13:51:27 crc kubenswrapper[4702]: I1125 13:51:27.534782 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-95b665d9-2w2fh" podUID="114a19ac-ef3a-4f27-8629-4faf56bb6943" containerName="controller-manager" containerID="cri-o://e994b359a7914e9ff3d40252fb13cad8312e29797145c0b3d8b72e4ae3b79c6f" gracePeriod=30 Nov 25 13:51:27 crc kubenswrapper[4702]: I1125 13:51:27.554598 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6ccdd5cd9-8swj9"] Nov 25 13:51:27 crc kubenswrapper[4702]: I1125 13:51:27.554804 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6ccdd5cd9-8swj9" podUID="e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c" containerName="route-controller-manager" containerID="cri-o://202057cb377ce8b78080795c951c2bd14cae0fdde47395222c4f5227f4cc8174" gracePeriod=30 Nov 25 13:51:27 crc kubenswrapper[4702]: I1125 13:51:27.959627 4702 generic.go:334] "Generic (PLEG): container finished" podID="114a19ac-ef3a-4f27-8629-4faf56bb6943" containerID="e994b359a7914e9ff3d40252fb13cad8312e29797145c0b3d8b72e4ae3b79c6f" exitCode=0 Nov 25 13:51:27 crc kubenswrapper[4702]: I1125 13:51:27.959699 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-95b665d9-2w2fh" event={"ID":"114a19ac-ef3a-4f27-8629-4faf56bb6943","Type":"ContainerDied","Data":"e994b359a7914e9ff3d40252fb13cad8312e29797145c0b3d8b72e4ae3b79c6f"} Nov 25 13:51:27 crc kubenswrapper[4702]: I1125 13:51:27.961055 4702 generic.go:334] "Generic (PLEG): container finished" podID="e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c" containerID="202057cb377ce8b78080795c951c2bd14cae0fdde47395222c4f5227f4cc8174" exitCode=0 Nov 25 13:51:27 crc kubenswrapper[4702]: I1125 13:51:27.961107 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6ccdd5cd9-8swj9" event={"ID":"e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c","Type":"ContainerDied","Data":"202057cb377ce8b78080795c951c2bd14cae0fdde47395222c4f5227f4cc8174"} Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.068393 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6ccdd5cd9-8swj9" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.153530 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-95b665d9-2w2fh" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.182731 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c-config\") pod \"e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c\" (UID: \"e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c\") " Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.182823 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c-client-ca\") pod \"e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c\" (UID: \"e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c\") " Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.182893 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bgz7\" (UniqueName: \"kubernetes.io/projected/e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c-kube-api-access-9bgz7\") pod \"e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c\" (UID: \"e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c\") " Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.182920 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c-serving-cert\") pod \"e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c\" (UID: \"e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c\") " Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.183980 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c-client-ca" (OuterVolumeSpecName: "client-ca") pod "e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c" (UID: "e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.183990 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c-config" (OuterVolumeSpecName: "config") pod "e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c" (UID: "e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.190305 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c" (UID: "e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.190350 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c-kube-api-access-9bgz7" (OuterVolumeSpecName: "kube-api-access-9bgz7") pod "e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c" (UID: "e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c"). InnerVolumeSpecName "kube-api-access-9bgz7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.284449 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/114a19ac-ef3a-4f27-8629-4faf56bb6943-config\") pod \"114a19ac-ef3a-4f27-8629-4faf56bb6943\" (UID: \"114a19ac-ef3a-4f27-8629-4faf56bb6943\") " Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.284970 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dh9k\" (UniqueName: \"kubernetes.io/projected/114a19ac-ef3a-4f27-8629-4faf56bb6943-kube-api-access-8dh9k\") pod \"114a19ac-ef3a-4f27-8629-4faf56bb6943\" (UID: \"114a19ac-ef3a-4f27-8629-4faf56bb6943\") " Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.284998 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/114a19ac-ef3a-4f27-8629-4faf56bb6943-client-ca\") pod \"114a19ac-ef3a-4f27-8629-4faf56bb6943\" (UID: \"114a19ac-ef3a-4f27-8629-4faf56bb6943\") " Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.285059 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/114a19ac-ef3a-4f27-8629-4faf56bb6943-serving-cert\") pod \"114a19ac-ef3a-4f27-8629-4faf56bb6943\" (UID: \"114a19ac-ef3a-4f27-8629-4faf56bb6943\") " Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.285111 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/114a19ac-ef3a-4f27-8629-4faf56bb6943-proxy-ca-bundles\") pod \"114a19ac-ef3a-4f27-8629-4faf56bb6943\" (UID: \"114a19ac-ef3a-4f27-8629-4faf56bb6943\") " Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.285532 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bgz7\" (UniqueName: \"kubernetes.io/projected/e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c-kube-api-access-9bgz7\") on node \"crc\" DevicePath \"\"" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.285547 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.285557 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c-config\") on node \"crc\" DevicePath \"\"" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.285569 4702 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.286051 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/114a19ac-ef3a-4f27-8629-4faf56bb6943-config" (OuterVolumeSpecName: "config") pod "114a19ac-ef3a-4f27-8629-4faf56bb6943" (UID: "114a19ac-ef3a-4f27-8629-4faf56bb6943"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.286173 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/114a19ac-ef3a-4f27-8629-4faf56bb6943-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "114a19ac-ef3a-4f27-8629-4faf56bb6943" (UID: "114a19ac-ef3a-4f27-8629-4faf56bb6943"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.286257 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/114a19ac-ef3a-4f27-8629-4faf56bb6943-client-ca" (OuterVolumeSpecName: "client-ca") pod "114a19ac-ef3a-4f27-8629-4faf56bb6943" (UID: "114a19ac-ef3a-4f27-8629-4faf56bb6943"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.288758 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/114a19ac-ef3a-4f27-8629-4faf56bb6943-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "114a19ac-ef3a-4f27-8629-4faf56bb6943" (UID: "114a19ac-ef3a-4f27-8629-4faf56bb6943"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.289607 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/114a19ac-ef3a-4f27-8629-4faf56bb6943-kube-api-access-8dh9k" (OuterVolumeSpecName: "kube-api-access-8dh9k") pod "114a19ac-ef3a-4f27-8629-4faf56bb6943" (UID: "114a19ac-ef3a-4f27-8629-4faf56bb6943"). InnerVolumeSpecName "kube-api-access-8dh9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.387520 4702 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/114a19ac-ef3a-4f27-8629-4faf56bb6943-config\") on node \"crc\" DevicePath \"\"" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.387582 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dh9k\" (UniqueName: \"kubernetes.io/projected/114a19ac-ef3a-4f27-8629-4faf56bb6943-kube-api-access-8dh9k\") on node \"crc\" DevicePath \"\"" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.387599 4702 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/114a19ac-ef3a-4f27-8629-4faf56bb6943-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.387610 4702 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/114a19ac-ef3a-4f27-8629-4faf56bb6943-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.387622 4702 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/114a19ac-ef3a-4f27-8629-4faf56bb6943-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.449408 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5fmgv" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.449585 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5fmgv" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.858174 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5567bbc789-5kpzt"] Nov 25 13:51:28 crc kubenswrapper[4702]: E1125 13:51:28.858433 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="114a19ac-ef3a-4f27-8629-4faf56bb6943" containerName="controller-manager" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.858446 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="114a19ac-ef3a-4f27-8629-4faf56bb6943" containerName="controller-manager" Nov 25 13:51:28 crc kubenswrapper[4702]: E1125 13:51:28.858457 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c" containerName="route-controller-manager" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.858462 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c" containerName="route-controller-manager" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.858585 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c" containerName="route-controller-manager" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.858599 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="114a19ac-ef3a-4f27-8629-4faf56bb6943" containerName="controller-manager" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.858992 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5567bbc789-5kpzt" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.863969 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7d86d9c56c-gsn5c"] Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.865021 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7d86d9c56c-gsn5c" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.869222 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7d86d9c56c-gsn5c"] Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.872476 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5567bbc789-5kpzt"] Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.968983 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-95b665d9-2w2fh" event={"ID":"114a19ac-ef3a-4f27-8629-4faf56bb6943","Type":"ContainerDied","Data":"ac1d7b6013969d14422d14bfcddb2506260c32c3bfd8b8a76d8bb45bcb7d8562"} Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.969068 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-95b665d9-2w2fh" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.969098 4702 scope.go:117] "RemoveContainer" containerID="e994b359a7914e9ff3d40252fb13cad8312e29797145c0b3d8b72e4ae3b79c6f" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.971211 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6ccdd5cd9-8swj9" event={"ID":"e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c","Type":"ContainerDied","Data":"51ec9b1e566bb2abc6fc8a89d4238004d5107dd7c47243798558b4b87d4069f9"} Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.971280 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6ccdd5cd9-8swj9" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.987472 4702 scope.go:117] "RemoveContainer" containerID="202057cb377ce8b78080795c951c2bd14cae0fdde47395222c4f5227f4cc8174" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.996463 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8298f7c7-6f80-4439-9885-c4d0025ffb50-serving-cert\") pod \"controller-manager-7d86d9c56c-gsn5c\" (UID: \"8298f7c7-6f80-4439-9885-c4d0025ffb50\") " pod="openshift-controller-manager/controller-manager-7d86d9c56c-gsn5c" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.996520 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71282f31-2d03-4b61-9d1f-70fd58c20592-serving-cert\") pod \"route-controller-manager-5567bbc789-5kpzt\" (UID: \"71282f31-2d03-4b61-9d1f-70fd58c20592\") " pod="openshift-route-controller-manager/route-controller-manager-5567bbc789-5kpzt" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.996549 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvl68\" (UniqueName: \"kubernetes.io/projected/71282f31-2d03-4b61-9d1f-70fd58c20592-kube-api-access-vvl68\") pod \"route-controller-manager-5567bbc789-5kpzt\" (UID: \"71282f31-2d03-4b61-9d1f-70fd58c20592\") " pod="openshift-route-controller-manager/route-controller-manager-5567bbc789-5kpzt" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.996626 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8298f7c7-6f80-4439-9885-c4d0025ffb50-client-ca\") pod \"controller-manager-7d86d9c56c-gsn5c\" (UID: \"8298f7c7-6f80-4439-9885-c4d0025ffb50\") " pod="openshift-controller-manager/controller-manager-7d86d9c56c-gsn5c" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.996668 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8298f7c7-6f80-4439-9885-c4d0025ffb50-proxy-ca-bundles\") pod \"controller-manager-7d86d9c56c-gsn5c\" (UID: \"8298f7c7-6f80-4439-9885-c4d0025ffb50\") " pod="openshift-controller-manager/controller-manager-7d86d9c56c-gsn5c" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.996693 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8298f7c7-6f80-4439-9885-c4d0025ffb50-config\") pod \"controller-manager-7d86d9c56c-gsn5c\" (UID: \"8298f7c7-6f80-4439-9885-c4d0025ffb50\") " pod="openshift-controller-manager/controller-manager-7d86d9c56c-gsn5c" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.996718 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71282f31-2d03-4b61-9d1f-70fd58c20592-config\") pod \"route-controller-manager-5567bbc789-5kpzt\" (UID: \"71282f31-2d03-4b61-9d1f-70fd58c20592\") " pod="openshift-route-controller-manager/route-controller-manager-5567bbc789-5kpzt" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.996744 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nz7l2\" (UniqueName: \"kubernetes.io/projected/8298f7c7-6f80-4439-9885-c4d0025ffb50-kube-api-access-nz7l2\") pod \"controller-manager-7d86d9c56c-gsn5c\" (UID: \"8298f7c7-6f80-4439-9885-c4d0025ffb50\") " pod="openshift-controller-manager/controller-manager-7d86d9c56c-gsn5c" Nov 25 13:51:28 crc kubenswrapper[4702]: I1125 13:51:28.996768 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/71282f31-2d03-4b61-9d1f-70fd58c20592-client-ca\") pod \"route-controller-manager-5567bbc789-5kpzt\" (UID: \"71282f31-2d03-4b61-9d1f-70fd58c20592\") " pod="openshift-route-controller-manager/route-controller-manager-5567bbc789-5kpzt" Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.020408 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6ccdd5cd9-8swj9"] Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.028109 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6ccdd5cd9-8swj9"] Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.033021 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-95b665d9-2w2fh"] Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.038402 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-95b665d9-2w2fh"] Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.054045 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-r28xx" Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.056961 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-r28xx" Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.096350 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-r28xx" Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.097506 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8298f7c7-6f80-4439-9885-c4d0025ffb50-client-ca\") pod \"controller-manager-7d86d9c56c-gsn5c\" (UID: \"8298f7c7-6f80-4439-9885-c4d0025ffb50\") " pod="openshift-controller-manager/controller-manager-7d86d9c56c-gsn5c" Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.097678 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8298f7c7-6f80-4439-9885-c4d0025ffb50-proxy-ca-bundles\") pod \"controller-manager-7d86d9c56c-gsn5c\" (UID: \"8298f7c7-6f80-4439-9885-c4d0025ffb50\") " pod="openshift-controller-manager/controller-manager-7d86d9c56c-gsn5c" Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.097802 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8298f7c7-6f80-4439-9885-c4d0025ffb50-config\") pod \"controller-manager-7d86d9c56c-gsn5c\" (UID: \"8298f7c7-6f80-4439-9885-c4d0025ffb50\") " pod="openshift-controller-manager/controller-manager-7d86d9c56c-gsn5c" Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.097886 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71282f31-2d03-4b61-9d1f-70fd58c20592-config\") pod \"route-controller-manager-5567bbc789-5kpzt\" (UID: \"71282f31-2d03-4b61-9d1f-70fd58c20592\") " pod="openshift-route-controller-manager/route-controller-manager-5567bbc789-5kpzt" Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.097968 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nz7l2\" (UniqueName: \"kubernetes.io/projected/8298f7c7-6f80-4439-9885-c4d0025ffb50-kube-api-access-nz7l2\") pod \"controller-manager-7d86d9c56c-gsn5c\" (UID: \"8298f7c7-6f80-4439-9885-c4d0025ffb50\") " pod="openshift-controller-manager/controller-manager-7d86d9c56c-gsn5c" Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.098046 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/71282f31-2d03-4b61-9d1f-70fd58c20592-client-ca\") pod \"route-controller-manager-5567bbc789-5kpzt\" (UID: \"71282f31-2d03-4b61-9d1f-70fd58c20592\") " pod="openshift-route-controller-manager/route-controller-manager-5567bbc789-5kpzt" Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.098129 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8298f7c7-6f80-4439-9885-c4d0025ffb50-serving-cert\") pod \"controller-manager-7d86d9c56c-gsn5c\" (UID: \"8298f7c7-6f80-4439-9885-c4d0025ffb50\") " pod="openshift-controller-manager/controller-manager-7d86d9c56c-gsn5c" Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.098211 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71282f31-2d03-4b61-9d1f-70fd58c20592-serving-cert\") pod \"route-controller-manager-5567bbc789-5kpzt\" (UID: \"71282f31-2d03-4b61-9d1f-70fd58c20592\") " pod="openshift-route-controller-manager/route-controller-manager-5567bbc789-5kpzt" Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.098304 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvl68\" (UniqueName: \"kubernetes.io/projected/71282f31-2d03-4b61-9d1f-70fd58c20592-kube-api-access-vvl68\") pod \"route-controller-manager-5567bbc789-5kpzt\" (UID: \"71282f31-2d03-4b61-9d1f-70fd58c20592\") " pod="openshift-route-controller-manager/route-controller-manager-5567bbc789-5kpzt" Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.098814 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8298f7c7-6f80-4439-9885-c4d0025ffb50-client-ca\") pod \"controller-manager-7d86d9c56c-gsn5c\" (UID: \"8298f7c7-6f80-4439-9885-c4d0025ffb50\") " pod="openshift-controller-manager/controller-manager-7d86d9c56c-gsn5c" Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.098829 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8298f7c7-6f80-4439-9885-c4d0025ffb50-proxy-ca-bundles\") pod \"controller-manager-7d86d9c56c-gsn5c\" (UID: \"8298f7c7-6f80-4439-9885-c4d0025ffb50\") " pod="openshift-controller-manager/controller-manager-7d86d9c56c-gsn5c" Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.099549 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/71282f31-2d03-4b61-9d1f-70fd58c20592-client-ca\") pod \"route-controller-manager-5567bbc789-5kpzt\" (UID: \"71282f31-2d03-4b61-9d1f-70fd58c20592\") " pod="openshift-route-controller-manager/route-controller-manager-5567bbc789-5kpzt" Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.099954 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8298f7c7-6f80-4439-9885-c4d0025ffb50-config\") pod \"controller-manager-7d86d9c56c-gsn5c\" (UID: \"8298f7c7-6f80-4439-9885-c4d0025ffb50\") " pod="openshift-controller-manager/controller-manager-7d86d9c56c-gsn5c" Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.100167 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71282f31-2d03-4b61-9d1f-70fd58c20592-config\") pod \"route-controller-manager-5567bbc789-5kpzt\" (UID: \"71282f31-2d03-4b61-9d1f-70fd58c20592\") " pod="openshift-route-controller-manager/route-controller-manager-5567bbc789-5kpzt" Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.104219 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8298f7c7-6f80-4439-9885-c4d0025ffb50-serving-cert\") pod \"controller-manager-7d86d9c56c-gsn5c\" (UID: \"8298f7c7-6f80-4439-9885-c4d0025ffb50\") " pod="openshift-controller-manager/controller-manager-7d86d9c56c-gsn5c" Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.112382 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71282f31-2d03-4b61-9d1f-70fd58c20592-serving-cert\") pod \"route-controller-manager-5567bbc789-5kpzt\" (UID: \"71282f31-2d03-4b61-9d1f-70fd58c20592\") " pod="openshift-route-controller-manager/route-controller-manager-5567bbc789-5kpzt" Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.118025 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nz7l2\" (UniqueName: \"kubernetes.io/projected/8298f7c7-6f80-4439-9885-c4d0025ffb50-kube-api-access-nz7l2\") pod \"controller-manager-7d86d9c56c-gsn5c\" (UID: \"8298f7c7-6f80-4439-9885-c4d0025ffb50\") " pod="openshift-controller-manager/controller-manager-7d86d9c56c-gsn5c" Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.118847 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvl68\" (UniqueName: \"kubernetes.io/projected/71282f31-2d03-4b61-9d1f-70fd58c20592-kube-api-access-vvl68\") pod \"route-controller-manager-5567bbc789-5kpzt\" (UID: \"71282f31-2d03-4b61-9d1f-70fd58c20592\") " pod="openshift-route-controller-manager/route-controller-manager-5567bbc789-5kpzt" Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.185712 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5567bbc789-5kpzt" Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.194034 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7d86d9c56c-gsn5c" Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.445731 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7d86d9c56c-gsn5c"] Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.505469 4702 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5fmgv" podUID="afc764af-1603-432c-9993-d0fc66fbb4fa" containerName="registry-server" probeResult="failure" output=< Nov 25 13:51:29 crc kubenswrapper[4702]: timeout: failed to connect service ":50051" within 1s Nov 25 13:51:29 crc kubenswrapper[4702]: > Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.606811 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5567bbc789-5kpzt"] Nov 25 13:51:29 crc kubenswrapper[4702]: W1125 13:51:29.616202 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71282f31_2d03_4b61_9d1f_70fd58c20592.slice/crio-fd803dbc36139c99ea8a5c97195f35c44faf542e64fcf7d44222aa5af149416e WatchSource:0}: Error finding container fd803dbc36139c99ea8a5c97195f35c44faf542e64fcf7d44222aa5af149416e: Status 404 returned error can't find the container with id fd803dbc36139c99ea8a5c97195f35c44faf542e64fcf7d44222aa5af149416e Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.810612 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="114a19ac-ef3a-4f27-8629-4faf56bb6943" path="/var/lib/kubelet/pods/114a19ac-ef3a-4f27-8629-4faf56bb6943/volumes" Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.811417 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c" path="/var/lib/kubelet/pods/e3cd004b-ae5e-4cfa-8496-1cf0fedacb8c/volumes" Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.979747 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7d86d9c56c-gsn5c" event={"ID":"8298f7c7-6f80-4439-9885-c4d0025ffb50","Type":"ContainerStarted","Data":"e84cf0be5ea7b04b1a71918711919f0fbd28b4fd4da6c3779373ea141e482a4f"} Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.980356 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7d86d9c56c-gsn5c" Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.980381 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7d86d9c56c-gsn5c" event={"ID":"8298f7c7-6f80-4439-9885-c4d0025ffb50","Type":"ContainerStarted","Data":"421d61cfec3fec091a2e28ce6f549816ee3876a1aae9e402dec24a76b01c1f2c"} Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.982298 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5567bbc789-5kpzt" event={"ID":"71282f31-2d03-4b61-9d1f-70fd58c20592","Type":"ContainerStarted","Data":"d5063e393fe736007af64f12345848d01bdc6111ec99fe684105e03de8220be1"} Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.982329 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5567bbc789-5kpzt" event={"ID":"71282f31-2d03-4b61-9d1f-70fd58c20592","Type":"ContainerStarted","Data":"fd803dbc36139c99ea8a5c97195f35c44faf542e64fcf7d44222aa5af149416e"} Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.983203 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5567bbc789-5kpzt" Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.984988 4702 patch_prober.go:28] interesting pod/route-controller-manager-5567bbc789-5kpzt container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.69:8443/healthz\": dial tcp 10.217.0.69:8443: connect: connection refused" start-of-body= Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.985040 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-5567bbc789-5kpzt" podUID="71282f31-2d03-4b61-9d1f-70fd58c20592" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.69:8443/healthz\": dial tcp 10.217.0.69:8443: connect: connection refused" Nov 25 13:51:29 crc kubenswrapper[4702]: I1125 13:51:29.989469 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7d86d9c56c-gsn5c" Nov 25 13:51:30 crc kubenswrapper[4702]: I1125 13:51:30.004873 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7d86d9c56c-gsn5c" podStartSLOduration=3.004847663 podStartE2EDuration="3.004847663s" podCreationTimestamp="2025-11-25 13:51:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:51:29.998712707 +0000 UTC m=+299.161363167" watchObservedRunningTime="2025-11-25 13:51:30.004847663 +0000 UTC m=+299.167498143" Nov 25 13:51:30 crc kubenswrapper[4702]: I1125 13:51:30.044112 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-r28xx" Nov 25 13:51:30 crc kubenswrapper[4702]: I1125 13:51:30.045477 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5567bbc789-5kpzt" podStartSLOduration=3.045462959 podStartE2EDuration="3.045462959s" podCreationTimestamp="2025-11-25 13:51:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:51:30.018420139 +0000 UTC m=+299.181070589" watchObservedRunningTime="2025-11-25 13:51:30.045462959 +0000 UTC m=+299.208113409" Nov 25 13:51:30 crc kubenswrapper[4702]: I1125 13:51:30.997997 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5567bbc789-5kpzt" Nov 25 13:51:31 crc kubenswrapper[4702]: I1125 13:51:31.236730 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xlvf4" Nov 25 13:51:31 crc kubenswrapper[4702]: I1125 13:51:31.236807 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xlvf4" Nov 25 13:51:31 crc kubenswrapper[4702]: I1125 13:51:31.285288 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xlvf4" Nov 25 13:51:31 crc kubenswrapper[4702]: I1125 13:51:31.528650 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nhp97" Nov 25 13:51:31 crc kubenswrapper[4702]: I1125 13:51:31.528707 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nhp97" Nov 25 13:51:31 crc kubenswrapper[4702]: I1125 13:51:31.570991 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nhp97" Nov 25 13:51:32 crc kubenswrapper[4702]: I1125 13:51:32.033208 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nhp97" Nov 25 13:51:32 crc kubenswrapper[4702]: I1125 13:51:32.033594 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xlvf4" Nov 25 13:51:38 crc kubenswrapper[4702]: I1125 13:51:38.491225 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5fmgv" Nov 25 13:51:38 crc kubenswrapper[4702]: I1125 13:51:38.533791 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5fmgv" Nov 25 13:51:39 crc kubenswrapper[4702]: I1125 13:51:39.288462 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-k8g4d" Nov 25 13:51:39 crc kubenswrapper[4702]: I1125 13:51:39.340848 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vlz75"] Nov 25 13:52:04 crc kubenswrapper[4702]: I1125 13:52:04.385253 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" podUID="a70d6c59-6b85-4950-8e30-1938da7a01e8" containerName="registry" containerID="cri-o://2c3c42ca85810566c17151205d72d3310991d375f0e4e80d8686edd5c6695b87" gracePeriod=30 Nov 25 13:52:04 crc kubenswrapper[4702]: I1125 13:52:04.866922 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:52:04 crc kubenswrapper[4702]: I1125 13:52:04.971567 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a70d6c59-6b85-4950-8e30-1938da7a01e8-installation-pull-secrets\") pod \"a70d6c59-6b85-4950-8e30-1938da7a01e8\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " Nov 25 13:52:04 crc kubenswrapper[4702]: I1125 13:52:04.971640 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a70d6c59-6b85-4950-8e30-1938da7a01e8-registry-tls\") pod \"a70d6c59-6b85-4950-8e30-1938da7a01e8\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " Nov 25 13:52:04 crc kubenswrapper[4702]: I1125 13:52:04.971674 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a70d6c59-6b85-4950-8e30-1938da7a01e8-ca-trust-extracted\") pod \"a70d6c59-6b85-4950-8e30-1938da7a01e8\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " Nov 25 13:52:04 crc kubenswrapper[4702]: I1125 13:52:04.972116 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a70d6c59-6b85-4950-8e30-1938da7a01e8-trusted-ca\") pod \"a70d6c59-6b85-4950-8e30-1938da7a01e8\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " Nov 25 13:52:04 crc kubenswrapper[4702]: I1125 13:52:04.972513 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"a70d6c59-6b85-4950-8e30-1938da7a01e8\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " Nov 25 13:52:04 crc kubenswrapper[4702]: I1125 13:52:04.972572 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a70d6c59-6b85-4950-8e30-1938da7a01e8-registry-certificates\") pod \"a70d6c59-6b85-4950-8e30-1938da7a01e8\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " Nov 25 13:52:04 crc kubenswrapper[4702]: I1125 13:52:04.972628 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a70d6c59-6b85-4950-8e30-1938da7a01e8-bound-sa-token\") pod \"a70d6c59-6b85-4950-8e30-1938da7a01e8\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " Nov 25 13:52:04 crc kubenswrapper[4702]: I1125 13:52:04.972659 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5829q\" (UniqueName: \"kubernetes.io/projected/a70d6c59-6b85-4950-8e30-1938da7a01e8-kube-api-access-5829q\") pod \"a70d6c59-6b85-4950-8e30-1938da7a01e8\" (UID: \"a70d6c59-6b85-4950-8e30-1938da7a01e8\") " Nov 25 13:52:04 crc kubenswrapper[4702]: I1125 13:52:04.974543 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a70d6c59-6b85-4950-8e30-1938da7a01e8-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a70d6c59-6b85-4950-8e30-1938da7a01e8" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:52:04 crc kubenswrapper[4702]: I1125 13:52:04.975189 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a70d6c59-6b85-4950-8e30-1938da7a01e8-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "a70d6c59-6b85-4950-8e30-1938da7a01e8" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:52:04 crc kubenswrapper[4702]: I1125 13:52:04.978902 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a70d6c59-6b85-4950-8e30-1938da7a01e8-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "a70d6c59-6b85-4950-8e30-1938da7a01e8" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:52:04 crc kubenswrapper[4702]: I1125 13:52:04.979100 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a70d6c59-6b85-4950-8e30-1938da7a01e8-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "a70d6c59-6b85-4950-8e30-1938da7a01e8" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:52:04 crc kubenswrapper[4702]: I1125 13:52:04.979511 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a70d6c59-6b85-4950-8e30-1938da7a01e8-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a70d6c59-6b85-4950-8e30-1938da7a01e8" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:52:04 crc kubenswrapper[4702]: I1125 13:52:04.986325 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a70d6c59-6b85-4950-8e30-1938da7a01e8-kube-api-access-5829q" (OuterVolumeSpecName: "kube-api-access-5829q") pod "a70d6c59-6b85-4950-8e30-1938da7a01e8" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8"). InnerVolumeSpecName "kube-api-access-5829q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:52:04 crc kubenswrapper[4702]: I1125 13:52:04.995841 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "a70d6c59-6b85-4950-8e30-1938da7a01e8" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 25 13:52:04 crc kubenswrapper[4702]: I1125 13:52:04.998733 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a70d6c59-6b85-4950-8e30-1938da7a01e8-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "a70d6c59-6b85-4950-8e30-1938da7a01e8" (UID: "a70d6c59-6b85-4950-8e30-1938da7a01e8"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:52:05 crc kubenswrapper[4702]: I1125 13:52:05.074568 4702 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a70d6c59-6b85-4950-8e30-1938da7a01e8-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 25 13:52:05 crc kubenswrapper[4702]: I1125 13:52:05.074633 4702 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a70d6c59-6b85-4950-8e30-1938da7a01e8-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 13:52:05 crc kubenswrapper[4702]: I1125 13:52:05.074645 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5829q\" (UniqueName: \"kubernetes.io/projected/a70d6c59-6b85-4950-8e30-1938da7a01e8-kube-api-access-5829q\") on node \"crc\" DevicePath \"\"" Nov 25 13:52:05 crc kubenswrapper[4702]: I1125 13:52:05.074659 4702 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a70d6c59-6b85-4950-8e30-1938da7a01e8-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 25 13:52:05 crc kubenswrapper[4702]: I1125 13:52:05.074675 4702 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a70d6c59-6b85-4950-8e30-1938da7a01e8-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 25 13:52:05 crc kubenswrapper[4702]: I1125 13:52:05.074691 4702 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a70d6c59-6b85-4950-8e30-1938da7a01e8-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 25 13:52:05 crc kubenswrapper[4702]: I1125 13:52:05.074704 4702 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a70d6c59-6b85-4950-8e30-1938da7a01e8-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 13:52:05 crc kubenswrapper[4702]: I1125 13:52:05.216432 4702 generic.go:334] "Generic (PLEG): container finished" podID="a70d6c59-6b85-4950-8e30-1938da7a01e8" containerID="2c3c42ca85810566c17151205d72d3310991d375f0e4e80d8686edd5c6695b87" exitCode=0 Nov 25 13:52:05 crc kubenswrapper[4702]: I1125 13:52:05.216476 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" event={"ID":"a70d6c59-6b85-4950-8e30-1938da7a01e8","Type":"ContainerDied","Data":"2c3c42ca85810566c17151205d72d3310991d375f0e4e80d8686edd5c6695b87"} Nov 25 13:52:05 crc kubenswrapper[4702]: I1125 13:52:05.216504 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" event={"ID":"a70d6c59-6b85-4950-8e30-1938da7a01e8","Type":"ContainerDied","Data":"a3443723263d837e8e18ee56c1ff63c829f60df2aeab2f735bf225cb08c91c98"} Nov 25 13:52:05 crc kubenswrapper[4702]: I1125 13:52:05.216522 4702 scope.go:117] "RemoveContainer" containerID="2c3c42ca85810566c17151205d72d3310991d375f0e4e80d8686edd5c6695b87" Nov 25 13:52:05 crc kubenswrapper[4702]: I1125 13:52:05.216531 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-vlz75" Nov 25 13:52:05 crc kubenswrapper[4702]: I1125 13:52:05.232908 4702 scope.go:117] "RemoveContainer" containerID="2c3c42ca85810566c17151205d72d3310991d375f0e4e80d8686edd5c6695b87" Nov 25 13:52:05 crc kubenswrapper[4702]: E1125 13:52:05.233441 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c3c42ca85810566c17151205d72d3310991d375f0e4e80d8686edd5c6695b87\": container with ID starting with 2c3c42ca85810566c17151205d72d3310991d375f0e4e80d8686edd5c6695b87 not found: ID does not exist" containerID="2c3c42ca85810566c17151205d72d3310991d375f0e4e80d8686edd5c6695b87" Nov 25 13:52:05 crc kubenswrapper[4702]: I1125 13:52:05.233479 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c3c42ca85810566c17151205d72d3310991d375f0e4e80d8686edd5c6695b87"} err="failed to get container status \"2c3c42ca85810566c17151205d72d3310991d375f0e4e80d8686edd5c6695b87\": rpc error: code = NotFound desc = could not find container \"2c3c42ca85810566c17151205d72d3310991d375f0e4e80d8686edd5c6695b87\": container with ID starting with 2c3c42ca85810566c17151205d72d3310991d375f0e4e80d8686edd5c6695b87 not found: ID does not exist" Nov 25 13:52:05 crc kubenswrapper[4702]: I1125 13:52:05.250019 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vlz75"] Nov 25 13:52:05 crc kubenswrapper[4702]: I1125 13:52:05.254255 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vlz75"] Nov 25 13:52:05 crc kubenswrapper[4702]: I1125 13:52:05.814044 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a70d6c59-6b85-4950-8e30-1938da7a01e8" path="/var/lib/kubelet/pods/a70d6c59-6b85-4950-8e30-1938da7a01e8/volumes" Nov 25 13:52:33 crc kubenswrapper[4702]: I1125 13:52:33.963272 4702 patch_prober.go:28] interesting pod/machine-config-daemon-qqr76 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 13:52:33 crc kubenswrapper[4702]: I1125 13:52:33.963819 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 13:53:03 crc kubenswrapper[4702]: I1125 13:53:03.962760 4702 patch_prober.go:28] interesting pod/machine-config-daemon-qqr76 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 13:53:03 crc kubenswrapper[4702]: I1125 13:53:03.963336 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 13:53:33 crc kubenswrapper[4702]: I1125 13:53:33.963639 4702 patch_prober.go:28] interesting pod/machine-config-daemon-qqr76 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 13:53:33 crc kubenswrapper[4702]: I1125 13:53:33.965008 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 13:53:33 crc kubenswrapper[4702]: I1125 13:53:33.965114 4702 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" Nov 25 13:53:33 crc kubenswrapper[4702]: I1125 13:53:33.966400 4702 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9283d384600217f266003d80798e4dd2a8e101df13b5073b74f9e20b9c61f5a6"} pod="openshift-machine-config-operator/machine-config-daemon-qqr76" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 13:53:33 crc kubenswrapper[4702]: I1125 13:53:33.966524 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" containerID="cri-o://9283d384600217f266003d80798e4dd2a8e101df13b5073b74f9e20b9c61f5a6" gracePeriod=600 Nov 25 13:53:34 crc kubenswrapper[4702]: I1125 13:53:34.748303 4702 generic.go:334] "Generic (PLEG): container finished" podID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerID="9283d384600217f266003d80798e4dd2a8e101df13b5073b74f9e20b9c61f5a6" exitCode=0 Nov 25 13:53:34 crc kubenswrapper[4702]: I1125 13:53:34.748432 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" event={"ID":"3c491818-b0bb-4d82-a031-96e2dfac8c27","Type":"ContainerDied","Data":"9283d384600217f266003d80798e4dd2a8e101df13b5073b74f9e20b9c61f5a6"} Nov 25 13:53:34 crc kubenswrapper[4702]: I1125 13:53:34.749424 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" event={"ID":"3c491818-b0bb-4d82-a031-96e2dfac8c27","Type":"ContainerStarted","Data":"a1697a0482710a529c531a772ee7967b9b00f72bbd70b22b58a644272dc2e190"} Nov 25 13:53:34 crc kubenswrapper[4702]: I1125 13:53:34.749460 4702 scope.go:117] "RemoveContainer" containerID="1a3eb592c71e1a4c84cf1bcb7b1c45a22de920b035dca9fbcf5b2486e75bfe7a" Nov 25 13:56:03 crc kubenswrapper[4702]: I1125 13:56:03.963668 4702 patch_prober.go:28] interesting pod/machine-config-daemon-qqr76 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 13:56:03 crc kubenswrapper[4702]: I1125 13:56:03.964391 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 13:56:33 crc kubenswrapper[4702]: I1125 13:56:33.963196 4702 patch_prober.go:28] interesting pod/machine-config-daemon-qqr76 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 13:56:33 crc kubenswrapper[4702]: I1125 13:56:33.964377 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.362452 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-p9c5v"] Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.363492 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="ovn-controller" containerID="cri-o://a5dfe629efca9398d6dcc541b5315c5b6d8e036cba1c94064883785b2cc7a1a4" gracePeriod=30 Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.363949 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="sbdb" containerID="cri-o://8700d12c2e236e704fc60136a51489d1d6890a83621480c23a6168d4a0f18f46" gracePeriod=30 Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.363996 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="nbdb" containerID="cri-o://c91a144aef48f19006041105594111c382420f9e6556a64580f6952990ef4db4" gracePeriod=30 Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.364043 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="northd" containerID="cri-o://7b2358ac7b25088c8d1b1cc5648a1fcbfd15211826d4c47f2c98839b9924138b" gracePeriod=30 Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.364083 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://a6d1bfac94a5fe6b382a9a3783250d6758603d00db9ce12ea44a3b54e6018103" gracePeriod=30 Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.364129 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="kube-rbac-proxy-node" containerID="cri-o://8a8fa25e9591d484f8dcb56974242924d4f31c7aef86b4376092842e57e1653c" gracePeriod=30 Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.364168 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="ovn-acl-logging" containerID="cri-o://9649c45bc796d60538bc1edc8f1239c7c5445de7b2f3dbf57dbe7df6ecab7d6c" gracePeriod=30 Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.403574 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="ovnkube-controller" containerID="cri-o://9e437f067d92fb24f70625cdd0e2a221164118fa19cd362de0a7cd5e71e88dd3" gracePeriod=30 Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.915013 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gd2xs_9e735a90-cf24-47de-b0e6-57d27a971b21/kube-multus/1.log" Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.916044 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gd2xs_9e735a90-cf24-47de-b0e6-57d27a971b21/kube-multus/0.log" Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.916083 4702 generic.go:334] "Generic (PLEG): container finished" podID="9e735a90-cf24-47de-b0e6-57d27a971b21" containerID="10ef8fba2e92332ed66f05f36bc57004f1d12c53689621106a3606f71d8aee9b" exitCode=2 Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.916142 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gd2xs" event={"ID":"9e735a90-cf24-47de-b0e6-57d27a971b21","Type":"ContainerDied","Data":"10ef8fba2e92332ed66f05f36bc57004f1d12c53689621106a3606f71d8aee9b"} Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.916183 4702 scope.go:117] "RemoveContainer" containerID="b88ff09a58bbf6e54d9cb8aa60ee993555f9fd94ecb8251046fd024722d60694" Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.916709 4702 scope.go:117] "RemoveContainer" containerID="10ef8fba2e92332ed66f05f36bc57004f1d12c53689621106a3606f71d8aee9b" Nov 25 13:56:50 crc kubenswrapper[4702]: E1125 13:56:50.916880 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-gd2xs_openshift-multus(9e735a90-cf24-47de-b0e6-57d27a971b21)\"" pod="openshift-multus/multus-gd2xs" podUID="9e735a90-cf24-47de-b0e6-57d27a971b21" Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.920033 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9c5v_3c8a3234-01d3-48a9-a5b4-acb46b4218dc/ovnkube-controller/2.log" Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.922608 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9c5v_3c8a3234-01d3-48a9-a5b4-acb46b4218dc/ovn-acl-logging/0.log" Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.923066 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9c5v_3c8a3234-01d3-48a9-a5b4-acb46b4218dc/ovn-controller/0.log" Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.923443 4702 generic.go:334] "Generic (PLEG): container finished" podID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerID="9e437f067d92fb24f70625cdd0e2a221164118fa19cd362de0a7cd5e71e88dd3" exitCode=0 Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.923462 4702 generic.go:334] "Generic (PLEG): container finished" podID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerID="8700d12c2e236e704fc60136a51489d1d6890a83621480c23a6168d4a0f18f46" exitCode=0 Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.923471 4702 generic.go:334] "Generic (PLEG): container finished" podID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerID="c91a144aef48f19006041105594111c382420f9e6556a64580f6952990ef4db4" exitCode=0 Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.923478 4702 generic.go:334] "Generic (PLEG): container finished" podID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerID="7b2358ac7b25088c8d1b1cc5648a1fcbfd15211826d4c47f2c98839b9924138b" exitCode=0 Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.923484 4702 generic.go:334] "Generic (PLEG): container finished" podID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerID="a6d1bfac94a5fe6b382a9a3783250d6758603d00db9ce12ea44a3b54e6018103" exitCode=0 Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.923490 4702 generic.go:334] "Generic (PLEG): container finished" podID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerID="8a8fa25e9591d484f8dcb56974242924d4f31c7aef86b4376092842e57e1653c" exitCode=0 Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.923499 4702 generic.go:334] "Generic (PLEG): container finished" podID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerID="9649c45bc796d60538bc1edc8f1239c7c5445de7b2f3dbf57dbe7df6ecab7d6c" exitCode=143 Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.923505 4702 generic.go:334] "Generic (PLEG): container finished" podID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerID="a5dfe629efca9398d6dcc541b5315c5b6d8e036cba1c94064883785b2cc7a1a4" exitCode=143 Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.923506 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" event={"ID":"3c8a3234-01d3-48a9-a5b4-acb46b4218dc","Type":"ContainerDied","Data":"9e437f067d92fb24f70625cdd0e2a221164118fa19cd362de0a7cd5e71e88dd3"} Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.923534 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" event={"ID":"3c8a3234-01d3-48a9-a5b4-acb46b4218dc","Type":"ContainerDied","Data":"8700d12c2e236e704fc60136a51489d1d6890a83621480c23a6168d4a0f18f46"} Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.923545 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" event={"ID":"3c8a3234-01d3-48a9-a5b4-acb46b4218dc","Type":"ContainerDied","Data":"c91a144aef48f19006041105594111c382420f9e6556a64580f6952990ef4db4"} Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.923553 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" event={"ID":"3c8a3234-01d3-48a9-a5b4-acb46b4218dc","Type":"ContainerDied","Data":"7b2358ac7b25088c8d1b1cc5648a1fcbfd15211826d4c47f2c98839b9924138b"} Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.923562 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" event={"ID":"3c8a3234-01d3-48a9-a5b4-acb46b4218dc","Type":"ContainerDied","Data":"a6d1bfac94a5fe6b382a9a3783250d6758603d00db9ce12ea44a3b54e6018103"} Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.923571 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" event={"ID":"3c8a3234-01d3-48a9-a5b4-acb46b4218dc","Type":"ContainerDied","Data":"8a8fa25e9591d484f8dcb56974242924d4f31c7aef86b4376092842e57e1653c"} Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.923580 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" event={"ID":"3c8a3234-01d3-48a9-a5b4-acb46b4218dc","Type":"ContainerDied","Data":"9649c45bc796d60538bc1edc8f1239c7c5445de7b2f3dbf57dbe7df6ecab7d6c"} Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.923589 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" event={"ID":"3c8a3234-01d3-48a9-a5b4-acb46b4218dc","Type":"ContainerDied","Data":"a5dfe629efca9398d6dcc541b5315c5b6d8e036cba1c94064883785b2cc7a1a4"} Nov 25 13:56:50 crc kubenswrapper[4702]: I1125 13:56:50.940848 4702 scope.go:117] "RemoveContainer" containerID="e042cb58071abb8cf4a711485af66c8fa3259aca650ac4a09cb7d66e893c1cfc" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.126038 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9c5v_3c8a3234-01d3-48a9-a5b4-acb46b4218dc/ovn-acl-logging/0.log" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.126666 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9c5v_3c8a3234-01d3-48a9-a5b4-acb46b4218dc/ovn-controller/0.log" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.127331 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.175535 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-x6plt"] Nov 25 13:56:51 crc kubenswrapper[4702]: E1125 13:56:51.175737 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="kube-rbac-proxy-node" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.175751 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="kube-rbac-proxy-node" Nov 25 13:56:51 crc kubenswrapper[4702]: E1125 13:56:51.175764 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="nbdb" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.175773 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="nbdb" Nov 25 13:56:51 crc kubenswrapper[4702]: E1125 13:56:51.175780 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="ovn-controller" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.175786 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="ovn-controller" Nov 25 13:56:51 crc kubenswrapper[4702]: E1125 13:56:51.175794 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="ovnkube-controller" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.175800 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="ovnkube-controller" Nov 25 13:56:51 crc kubenswrapper[4702]: E1125 13:56:51.175809 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="ovnkube-controller" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.175815 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="ovnkube-controller" Nov 25 13:56:51 crc kubenswrapper[4702]: E1125 13:56:51.175825 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="northd" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.175831 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="northd" Nov 25 13:56:51 crc kubenswrapper[4702]: E1125 13:56:51.175838 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="ovn-acl-logging" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.175844 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="ovn-acl-logging" Nov 25 13:56:51 crc kubenswrapper[4702]: E1125 13:56:51.175852 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.175857 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 13:56:51 crc kubenswrapper[4702]: E1125 13:56:51.175866 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="kubecfg-setup" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.175872 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="kubecfg-setup" Nov 25 13:56:51 crc kubenswrapper[4702]: E1125 13:56:51.175884 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a70d6c59-6b85-4950-8e30-1938da7a01e8" containerName="registry" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.175891 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="a70d6c59-6b85-4950-8e30-1938da7a01e8" containerName="registry" Nov 25 13:56:51 crc kubenswrapper[4702]: E1125 13:56:51.175899 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="sbdb" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.175906 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="sbdb" Nov 25 13:56:51 crc kubenswrapper[4702]: E1125 13:56:51.175914 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="ovnkube-controller" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.175920 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="ovnkube-controller" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.176012 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="nbdb" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.176023 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="ovn-controller" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.176032 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="a70d6c59-6b85-4950-8e30-1938da7a01e8" containerName="registry" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.176039 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="ovnkube-controller" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.176048 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="ovnkube-controller" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.176057 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="ovn-acl-logging" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.176064 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="sbdb" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.176073 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="kube-rbac-proxy-node" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.176082 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.176089 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="northd" Nov 25 13:56:51 crc kubenswrapper[4702]: E1125 13:56:51.176173 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="ovnkube-controller" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.176181 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="ovnkube-controller" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.176286 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="ovnkube-controller" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.176296 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" containerName="ovnkube-controller" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.180287 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.235421 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-node-log\") pod \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.235472 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-etc-openvswitch\") pod \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.235501 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-var-lib-cni-networks-ovn-kubernetes\") pod \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.235519 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-run-systemd\") pod \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.235540 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-slash\") pod \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.235548 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-node-log" (OuterVolumeSpecName: "node-log") pod "3c8a3234-01d3-48a9-a5b4-acb46b4218dc" (UID: "3c8a3234-01d3-48a9-a5b4-acb46b4218dc"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.235564 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-ovn-node-metrics-cert\") pod \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.235588 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "3c8a3234-01d3-48a9-a5b4-acb46b4218dc" (UID: "3c8a3234-01d3-48a9-a5b4-acb46b4218dc"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.235594 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-kubelet\") pod \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.235609 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-slash" (OuterVolumeSpecName: "host-slash") pod "3c8a3234-01d3-48a9-a5b4-acb46b4218dc" (UID: "3c8a3234-01d3-48a9-a5b4-acb46b4218dc"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.235613 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-run-openvswitch\") pod \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.235609 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "3c8a3234-01d3-48a9-a5b4-acb46b4218dc" (UID: "3c8a3234-01d3-48a9-a5b4-acb46b4218dc"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.235652 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-ovnkube-config\") pod \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.235636 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "3c8a3234-01d3-48a9-a5b4-acb46b4218dc" (UID: "3c8a3234-01d3-48a9-a5b4-acb46b4218dc"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.235689 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "3c8a3234-01d3-48a9-a5b4-acb46b4218dc" (UID: "3c8a3234-01d3-48a9-a5b4-acb46b4218dc"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.235675 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-var-lib-openvswitch\") pod \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.235749 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-run-ovn-kubernetes\") pod \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.235802 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-log-socket\") pod \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.235831 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-systemd-units\") pod \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.235857 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-cni-bin\") pod \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.235878 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-run-ovn\") pod \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.235917 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-ovnkube-script-lib\") pod \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.235947 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "3c8a3234-01d3-48a9-a5b4-acb46b4218dc" (UID: "3c8a3234-01d3-48a9-a5b4-acb46b4218dc"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.235963 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-env-overrides\") pod \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.235973 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "3c8a3234-01d3-48a9-a5b4-acb46b4218dc" (UID: "3c8a3234-01d3-48a9-a5b4-acb46b4218dc"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.235995 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "3c8a3234-01d3-48a9-a5b4-acb46b4218dc" (UID: "3c8a3234-01d3-48a9-a5b4-acb46b4218dc"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.236000 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdztt\" (UniqueName: \"kubernetes.io/projected/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-kube-api-access-hdztt\") pod \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.236015 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-log-socket" (OuterVolumeSpecName: "log-socket") pod "3c8a3234-01d3-48a9-a5b4-acb46b4218dc" (UID: "3c8a3234-01d3-48a9-a5b4-acb46b4218dc"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.236025 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-run-netns\") pod \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.236076 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-cni-netd\") pod \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\" (UID: \"3c8a3234-01d3-48a9-a5b4-acb46b4218dc\") " Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.236301 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "3c8a3234-01d3-48a9-a5b4-acb46b4218dc" (UID: "3c8a3234-01d3-48a9-a5b4-acb46b4218dc"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.236300 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "3c8a3234-01d3-48a9-a5b4-acb46b4218dc" (UID: "3c8a3234-01d3-48a9-a5b4-acb46b4218dc"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.236329 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "3c8a3234-01d3-48a9-a5b4-acb46b4218dc" (UID: "3c8a3234-01d3-48a9-a5b4-acb46b4218dc"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.236356 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "3c8a3234-01d3-48a9-a5b4-acb46b4218dc" (UID: "3c8a3234-01d3-48a9-a5b4-acb46b4218dc"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.236380 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "3c8a3234-01d3-48a9-a5b4-acb46b4218dc" (UID: "3c8a3234-01d3-48a9-a5b4-acb46b4218dc"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.236456 4702 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.236473 4702 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.236483 4702 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-node-log\") on node \"crc\" DevicePath \"\"" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.236493 4702 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.236502 4702 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.236511 4702 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-slash\") on node \"crc\" DevicePath \"\"" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.236519 4702 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.236527 4702 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.236535 4702 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.236544 4702 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.236552 4702 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-log-socket\") on node \"crc\" DevicePath \"\"" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.236561 4702 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.236568 4702 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.236576 4702 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.236584 4702 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.236803 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "3c8a3234-01d3-48a9-a5b4-acb46b4218dc" (UID: "3c8a3234-01d3-48a9-a5b4-acb46b4218dc"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.236824 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "3c8a3234-01d3-48a9-a5b4-acb46b4218dc" (UID: "3c8a3234-01d3-48a9-a5b4-acb46b4218dc"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.241676 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "3c8a3234-01d3-48a9-a5b4-acb46b4218dc" (UID: "3c8a3234-01d3-48a9-a5b4-acb46b4218dc"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.242470 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-kube-api-access-hdztt" (OuterVolumeSpecName: "kube-api-access-hdztt") pod "3c8a3234-01d3-48a9-a5b4-acb46b4218dc" (UID: "3c8a3234-01d3-48a9-a5b4-acb46b4218dc"). InnerVolumeSpecName "kube-api-access-hdztt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.250560 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "3c8a3234-01d3-48a9-a5b4-acb46b4218dc" (UID: "3c8a3234-01d3-48a9-a5b4-acb46b4218dc"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.337657 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-etc-openvswitch\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.337716 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/003bbdd5-ccda-41df-9490-74fda500e653-env-overrides\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.337864 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-log-socket\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.338008 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/003bbdd5-ccda-41df-9490-74fda500e653-ovn-node-metrics-cert\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.338034 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-run-systemd\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.338068 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-run-openvswitch\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.338092 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-node-log\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.338115 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-systemd-units\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.338134 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-host-cni-bin\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.338190 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-host-cni-netd\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.338275 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-host-run-netns\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.338323 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qs7v\" (UniqueName: \"kubernetes.io/projected/003bbdd5-ccda-41df-9490-74fda500e653-kube-api-access-5qs7v\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.338345 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-var-lib-openvswitch\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.338370 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-host-slash\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.338420 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/003bbdd5-ccda-41df-9490-74fda500e653-ovnkube-script-lib\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.338480 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.338545 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-host-kubelet\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.338577 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-run-ovn\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.338597 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/003bbdd5-ccda-41df-9490-74fda500e653-ovnkube-config\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.338622 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-host-run-ovn-kubernetes\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.338682 4702 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.338696 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdztt\" (UniqueName: \"kubernetes.io/projected/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-kube-api-access-hdztt\") on node \"crc\" DevicePath \"\"" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.338706 4702 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.338718 4702 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.338729 4702 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3c8a3234-01d3-48a9-a5b4-acb46b4218dc-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.440240 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-host-cni-netd\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.440320 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-host-run-netns\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.440347 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qs7v\" (UniqueName: \"kubernetes.io/projected/003bbdd5-ccda-41df-9490-74fda500e653-kube-api-access-5qs7v\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.440368 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-var-lib-openvswitch\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.440389 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-host-slash\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.440420 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/003bbdd5-ccda-41df-9490-74fda500e653-ovnkube-script-lib\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.440451 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.440483 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-host-kubelet\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.440505 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-run-ovn\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.440524 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/003bbdd5-ccda-41df-9490-74fda500e653-ovnkube-config\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.440545 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-host-run-ovn-kubernetes\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.440573 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-etc-openvswitch\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.440591 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/003bbdd5-ccda-41df-9490-74fda500e653-env-overrides\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.440612 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-log-socket\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.440638 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/003bbdd5-ccda-41df-9490-74fda500e653-ovn-node-metrics-cert\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.440652 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-run-systemd\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.440675 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-run-openvswitch\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.440706 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-node-log\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.440723 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-systemd-units\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.440737 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-host-cni-bin\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.440798 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-host-cni-bin\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.440836 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-host-cni-netd\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.440858 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-host-run-netns\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.441148 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-var-lib-openvswitch\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.441176 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-host-slash\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.441917 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/003bbdd5-ccda-41df-9490-74fda500e653-ovnkube-script-lib\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.441968 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.442001 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-host-kubelet\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.442030 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-run-ovn\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.442336 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-run-openvswitch\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.442465 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-run-systemd\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.442503 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-etc-openvswitch\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.442514 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/003bbdd5-ccda-41df-9490-74fda500e653-ovnkube-config\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.442534 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-host-run-ovn-kubernetes\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.442558 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-node-log\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.442569 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-systemd-units\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.442605 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/003bbdd5-ccda-41df-9490-74fda500e653-log-socket\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.442970 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/003bbdd5-ccda-41df-9490-74fda500e653-env-overrides\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.445603 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/003bbdd5-ccda-41df-9490-74fda500e653-ovn-node-metrics-cert\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.456330 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qs7v\" (UniqueName: \"kubernetes.io/projected/003bbdd5-ccda-41df-9490-74fda500e653-kube-api-access-5qs7v\") pod \"ovnkube-node-x6plt\" (UID: \"003bbdd5-ccda-41df-9490-74fda500e653\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.496606 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.931055 4702 generic.go:334] "Generic (PLEG): container finished" podID="003bbdd5-ccda-41df-9490-74fda500e653" containerID="728476f319efd926e916254e9eeb5c20edd3fff53a27985e0e9fbae5f4f0c322" exitCode=0 Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.931177 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" event={"ID":"003bbdd5-ccda-41df-9490-74fda500e653","Type":"ContainerDied","Data":"728476f319efd926e916254e9eeb5c20edd3fff53a27985e0e9fbae5f4f0c322"} Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.931490 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" event={"ID":"003bbdd5-ccda-41df-9490-74fda500e653","Type":"ContainerStarted","Data":"ef6ffe6fd08e0ed671c7ea90c19e1b847c3912791b69747617e4cfb8c0141d85"} Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.934502 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gd2xs_9e735a90-cf24-47de-b0e6-57d27a971b21/kube-multus/1.log" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.948266 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9c5v_3c8a3234-01d3-48a9-a5b4-acb46b4218dc/ovn-acl-logging/0.log" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.948855 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p9c5v_3c8a3234-01d3-48a9-a5b4-acb46b4218dc/ovn-controller/0.log" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.949329 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" event={"ID":"3c8a3234-01d3-48a9-a5b4-acb46b4218dc","Type":"ContainerDied","Data":"7732801e549701b7f73321af4fb72c468f45cae12714478a77d6aeacada97801"} Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.949376 4702 scope.go:117] "RemoveContainer" containerID="9e437f067d92fb24f70625cdd0e2a221164118fa19cd362de0a7cd5e71e88dd3" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.949420 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-p9c5v" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.970626 4702 scope.go:117] "RemoveContainer" containerID="8700d12c2e236e704fc60136a51489d1d6890a83621480c23a6168d4a0f18f46" Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.991673 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-p9c5v"] Nov 25 13:56:51 crc kubenswrapper[4702]: I1125 13:56:51.994561 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-p9c5v"] Nov 25 13:56:52 crc kubenswrapper[4702]: I1125 13:56:52.004521 4702 scope.go:117] "RemoveContainer" containerID="c91a144aef48f19006041105594111c382420f9e6556a64580f6952990ef4db4" Nov 25 13:56:52 crc kubenswrapper[4702]: I1125 13:56:52.023477 4702 scope.go:117] "RemoveContainer" containerID="7b2358ac7b25088c8d1b1cc5648a1fcbfd15211826d4c47f2c98839b9924138b" Nov 25 13:56:52 crc kubenswrapper[4702]: I1125 13:56:52.037920 4702 scope.go:117] "RemoveContainer" containerID="a6d1bfac94a5fe6b382a9a3783250d6758603d00db9ce12ea44a3b54e6018103" Nov 25 13:56:52 crc kubenswrapper[4702]: I1125 13:56:52.053883 4702 scope.go:117] "RemoveContainer" containerID="8a8fa25e9591d484f8dcb56974242924d4f31c7aef86b4376092842e57e1653c" Nov 25 13:56:52 crc kubenswrapper[4702]: I1125 13:56:52.066466 4702 scope.go:117] "RemoveContainer" containerID="9649c45bc796d60538bc1edc8f1239c7c5445de7b2f3dbf57dbe7df6ecab7d6c" Nov 25 13:56:52 crc kubenswrapper[4702]: I1125 13:56:52.081618 4702 scope.go:117] "RemoveContainer" containerID="a5dfe629efca9398d6dcc541b5315c5b6d8e036cba1c94064883785b2cc7a1a4" Nov 25 13:56:52 crc kubenswrapper[4702]: I1125 13:56:52.099366 4702 scope.go:117] "RemoveContainer" containerID="d3635f635f53b37bd64bf6f2e8f22ee480eaeb0003592ba9a5b4947124ea6ea8" Nov 25 13:56:52 crc kubenswrapper[4702]: I1125 13:56:52.958325 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" event={"ID":"003bbdd5-ccda-41df-9490-74fda500e653","Type":"ContainerStarted","Data":"41f3fed6fa7c456f79d881b1fec739055aae20ed36118e9dbcf377f2e8acd3bf"} Nov 25 13:56:52 crc kubenswrapper[4702]: I1125 13:56:52.958653 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" event={"ID":"003bbdd5-ccda-41df-9490-74fda500e653","Type":"ContainerStarted","Data":"b361d4ba640e78a1808732988b3bc0164d818401e05b491217e0a2b6db38cbf4"} Nov 25 13:56:52 crc kubenswrapper[4702]: I1125 13:56:52.958669 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" event={"ID":"003bbdd5-ccda-41df-9490-74fda500e653","Type":"ContainerStarted","Data":"53c2f1d7ecdd353f559ed5e7394caa12c7dd45ee3f7b2afbd0869f1780060326"} Nov 25 13:56:53 crc kubenswrapper[4702]: I1125 13:56:53.808193 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c8a3234-01d3-48a9-a5b4-acb46b4218dc" path="/var/lib/kubelet/pods/3c8a3234-01d3-48a9-a5b4-acb46b4218dc/volumes" Nov 25 13:56:53 crc kubenswrapper[4702]: I1125 13:56:53.966786 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" event={"ID":"003bbdd5-ccda-41df-9490-74fda500e653","Type":"ContainerStarted","Data":"b7244b9f9229a35c1be51ccf5136d938f5d40e83823d2ceefe60e9f113fc27e8"} Nov 25 13:56:53 crc kubenswrapper[4702]: I1125 13:56:53.967399 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" event={"ID":"003bbdd5-ccda-41df-9490-74fda500e653","Type":"ContainerStarted","Data":"d285bb590b0005f36e2b6d0531b05d3d7f320e9361a3644edc430dd593df1ff8"} Nov 25 13:56:54 crc kubenswrapper[4702]: I1125 13:56:54.975849 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" event={"ID":"003bbdd5-ccda-41df-9490-74fda500e653","Type":"ContainerStarted","Data":"06a776598c05bd18724105c2b5b1cd9bf751fc6ec44ac57a972067037cc47ff1"} Nov 25 13:56:56 crc kubenswrapper[4702]: I1125 13:56:56.990596 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" event={"ID":"003bbdd5-ccda-41df-9490-74fda500e653","Type":"ContainerStarted","Data":"2b57d0424402b9edb0736f27bb54a9c5ca120b045a78e9b93d8d63a6454618bb"} Nov 25 13:56:59 crc kubenswrapper[4702]: I1125 13:56:59.005978 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" event={"ID":"003bbdd5-ccda-41df-9490-74fda500e653","Type":"ContainerStarted","Data":"cabf294515d53a508076b6bfc821c604aaa9b7a97018bc7034e4d94044cea75e"} Nov 25 13:56:59 crc kubenswrapper[4702]: I1125 13:56:59.006791 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:59 crc kubenswrapper[4702]: I1125 13:56:59.006808 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:59 crc kubenswrapper[4702]: I1125 13:56:59.006822 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:59 crc kubenswrapper[4702]: I1125 13:56:59.050852 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" podStartSLOduration=8.050825219 podStartE2EDuration="8.050825219s" podCreationTimestamp="2025-11-25 13:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 13:56:59.048908848 +0000 UTC m=+628.211559298" watchObservedRunningTime="2025-11-25 13:56:59.050825219 +0000 UTC m=+628.213475679" Nov 25 13:56:59 crc kubenswrapper[4702]: I1125 13:56:59.054224 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:56:59 crc kubenswrapper[4702]: I1125 13:56:59.058073 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:57:03 crc kubenswrapper[4702]: I1125 13:57:03.963569 4702 patch_prober.go:28] interesting pod/machine-config-daemon-qqr76 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 13:57:03 crc kubenswrapper[4702]: I1125 13:57:03.964608 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 13:57:03 crc kubenswrapper[4702]: I1125 13:57:03.964703 4702 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" Nov 25 13:57:03 crc kubenswrapper[4702]: I1125 13:57:03.965973 4702 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a1697a0482710a529c531a772ee7967b9b00f72bbd70b22b58a644272dc2e190"} pod="openshift-machine-config-operator/machine-config-daemon-qqr76" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 13:57:03 crc kubenswrapper[4702]: I1125 13:57:03.966132 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" containerID="cri-o://a1697a0482710a529c531a772ee7967b9b00f72bbd70b22b58a644272dc2e190" gracePeriod=600 Nov 25 13:57:04 crc kubenswrapper[4702]: I1125 13:57:04.801081 4702 scope.go:117] "RemoveContainer" containerID="10ef8fba2e92332ed66f05f36bc57004f1d12c53689621106a3606f71d8aee9b" Nov 25 13:57:05 crc kubenswrapper[4702]: I1125 13:57:05.042561 4702 generic.go:334] "Generic (PLEG): container finished" podID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerID="a1697a0482710a529c531a772ee7967b9b00f72bbd70b22b58a644272dc2e190" exitCode=0 Nov 25 13:57:05 crc kubenswrapper[4702]: I1125 13:57:05.042682 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" event={"ID":"3c491818-b0bb-4d82-a031-96e2dfac8c27","Type":"ContainerDied","Data":"a1697a0482710a529c531a772ee7967b9b00f72bbd70b22b58a644272dc2e190"} Nov 25 13:57:05 crc kubenswrapper[4702]: I1125 13:57:05.043159 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" event={"ID":"3c491818-b0bb-4d82-a031-96e2dfac8c27","Type":"ContainerStarted","Data":"63ddfd96ea126ce49009797997159362938b731d35159663d233bd6f5c4e30d7"} Nov 25 13:57:05 crc kubenswrapper[4702]: I1125 13:57:05.043189 4702 scope.go:117] "RemoveContainer" containerID="9283d384600217f266003d80798e4dd2a8e101df13b5073b74f9e20b9c61f5a6" Nov 25 13:57:06 crc kubenswrapper[4702]: I1125 13:57:06.056828 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gd2xs_9e735a90-cf24-47de-b0e6-57d27a971b21/kube-multus/1.log" Nov 25 13:57:06 crc kubenswrapper[4702]: I1125 13:57:06.057395 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gd2xs" event={"ID":"9e735a90-cf24-47de-b0e6-57d27a971b21","Type":"ContainerStarted","Data":"ef3c2790db5a897d7aef79abe06e6344c35715ca45bf5411e60fa83bff69e220"} Nov 25 13:57:14 crc kubenswrapper[4702]: I1125 13:57:14.912900 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7"] Nov 25 13:57:14 crc kubenswrapper[4702]: I1125 13:57:14.915938 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7" Nov 25 13:57:14 crc kubenswrapper[4702]: I1125 13:57:14.918467 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 25 13:57:14 crc kubenswrapper[4702]: I1125 13:57:14.936856 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7"] Nov 25 13:57:15 crc kubenswrapper[4702]: I1125 13:57:15.070986 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/97dcf135-5184-42db-b23e-b39dcbe81c99-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7\" (UID: \"97dcf135-5184-42db-b23e-b39dcbe81c99\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7" Nov 25 13:57:15 crc kubenswrapper[4702]: I1125 13:57:15.071109 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2pzt\" (UniqueName: \"kubernetes.io/projected/97dcf135-5184-42db-b23e-b39dcbe81c99-kube-api-access-j2pzt\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7\" (UID: \"97dcf135-5184-42db-b23e-b39dcbe81c99\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7" Nov 25 13:57:15 crc kubenswrapper[4702]: I1125 13:57:15.071160 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/97dcf135-5184-42db-b23e-b39dcbe81c99-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7\" (UID: \"97dcf135-5184-42db-b23e-b39dcbe81c99\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7" Nov 25 13:57:15 crc kubenswrapper[4702]: I1125 13:57:15.172516 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/97dcf135-5184-42db-b23e-b39dcbe81c99-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7\" (UID: \"97dcf135-5184-42db-b23e-b39dcbe81c99\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7" Nov 25 13:57:15 crc kubenswrapper[4702]: I1125 13:57:15.173111 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2pzt\" (UniqueName: \"kubernetes.io/projected/97dcf135-5184-42db-b23e-b39dcbe81c99-kube-api-access-j2pzt\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7\" (UID: \"97dcf135-5184-42db-b23e-b39dcbe81c99\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7" Nov 25 13:57:15 crc kubenswrapper[4702]: I1125 13:57:15.173036 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/97dcf135-5184-42db-b23e-b39dcbe81c99-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7\" (UID: \"97dcf135-5184-42db-b23e-b39dcbe81c99\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7" Nov 25 13:57:15 crc kubenswrapper[4702]: I1125 13:57:15.173146 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/97dcf135-5184-42db-b23e-b39dcbe81c99-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7\" (UID: \"97dcf135-5184-42db-b23e-b39dcbe81c99\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7" Nov 25 13:57:15 crc kubenswrapper[4702]: I1125 13:57:15.173473 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/97dcf135-5184-42db-b23e-b39dcbe81c99-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7\" (UID: \"97dcf135-5184-42db-b23e-b39dcbe81c99\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7" Nov 25 13:57:15 crc kubenswrapper[4702]: I1125 13:57:15.195642 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2pzt\" (UniqueName: \"kubernetes.io/projected/97dcf135-5184-42db-b23e-b39dcbe81c99-kube-api-access-j2pzt\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7\" (UID: \"97dcf135-5184-42db-b23e-b39dcbe81c99\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7" Nov 25 13:57:15 crc kubenswrapper[4702]: I1125 13:57:15.233107 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7" Nov 25 13:57:15 crc kubenswrapper[4702]: I1125 13:57:15.691719 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7"] Nov 25 13:57:16 crc kubenswrapper[4702]: I1125 13:57:16.116894 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7" event={"ID":"97dcf135-5184-42db-b23e-b39dcbe81c99","Type":"ContainerStarted","Data":"2d182a04bfa8da7152d391481abc28d6c4d55919385c412cb52b35bb90fb801b"} Nov 25 13:57:16 crc kubenswrapper[4702]: I1125 13:57:16.116961 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7" event={"ID":"97dcf135-5184-42db-b23e-b39dcbe81c99","Type":"ContainerStarted","Data":"b1b45405c2ff33900b91520d291888a0823b1a3412db8369dbb0bb37b166cde0"} Nov 25 13:57:17 crc kubenswrapper[4702]: I1125 13:57:17.133065 4702 generic.go:334] "Generic (PLEG): container finished" podID="97dcf135-5184-42db-b23e-b39dcbe81c99" containerID="2d182a04bfa8da7152d391481abc28d6c4d55919385c412cb52b35bb90fb801b" exitCode=0 Nov 25 13:57:17 crc kubenswrapper[4702]: I1125 13:57:17.133124 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7" event={"ID":"97dcf135-5184-42db-b23e-b39dcbe81c99","Type":"ContainerDied","Data":"2d182a04bfa8da7152d391481abc28d6c4d55919385c412cb52b35bb90fb801b"} Nov 25 13:57:17 crc kubenswrapper[4702]: I1125 13:57:17.135177 4702 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 13:57:19 crc kubenswrapper[4702]: I1125 13:57:19.157635 4702 generic.go:334] "Generic (PLEG): container finished" podID="97dcf135-5184-42db-b23e-b39dcbe81c99" containerID="9ce2a1377827ff6bb666ba9a515a55bb5e09a51877cb929fffd262b246e251ce" exitCode=0 Nov 25 13:57:19 crc kubenswrapper[4702]: I1125 13:57:19.157698 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7" event={"ID":"97dcf135-5184-42db-b23e-b39dcbe81c99","Type":"ContainerDied","Data":"9ce2a1377827ff6bb666ba9a515a55bb5e09a51877cb929fffd262b246e251ce"} Nov 25 13:57:20 crc kubenswrapper[4702]: I1125 13:57:20.962699 4702 generic.go:334] "Generic (PLEG): container finished" podID="97dcf135-5184-42db-b23e-b39dcbe81c99" containerID="a3d163627ef2d14ebd084e367328234acdcc47bfef31a097ca7fa22ebed07af8" exitCode=0 Nov 25 13:57:20 crc kubenswrapper[4702]: I1125 13:57:20.962781 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7" event={"ID":"97dcf135-5184-42db-b23e-b39dcbe81c99","Type":"ContainerDied","Data":"a3d163627ef2d14ebd084e367328234acdcc47bfef31a097ca7fa22ebed07af8"} Nov 25 13:57:21 crc kubenswrapper[4702]: I1125 13:57:21.522842 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-x6plt" Nov 25 13:57:22 crc kubenswrapper[4702]: I1125 13:57:22.197812 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7" Nov 25 13:57:22 crc kubenswrapper[4702]: I1125 13:57:22.274900 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2pzt\" (UniqueName: \"kubernetes.io/projected/97dcf135-5184-42db-b23e-b39dcbe81c99-kube-api-access-j2pzt\") pod \"97dcf135-5184-42db-b23e-b39dcbe81c99\" (UID: \"97dcf135-5184-42db-b23e-b39dcbe81c99\") " Nov 25 13:57:22 crc kubenswrapper[4702]: I1125 13:57:22.276204 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/97dcf135-5184-42db-b23e-b39dcbe81c99-bundle\") pod \"97dcf135-5184-42db-b23e-b39dcbe81c99\" (UID: \"97dcf135-5184-42db-b23e-b39dcbe81c99\") " Nov 25 13:57:22 crc kubenswrapper[4702]: I1125 13:57:22.276427 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/97dcf135-5184-42db-b23e-b39dcbe81c99-util\") pod \"97dcf135-5184-42db-b23e-b39dcbe81c99\" (UID: \"97dcf135-5184-42db-b23e-b39dcbe81c99\") " Nov 25 13:57:22 crc kubenswrapper[4702]: I1125 13:57:22.277630 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97dcf135-5184-42db-b23e-b39dcbe81c99-bundle" (OuterVolumeSpecName: "bundle") pod "97dcf135-5184-42db-b23e-b39dcbe81c99" (UID: "97dcf135-5184-42db-b23e-b39dcbe81c99"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:57:22 crc kubenswrapper[4702]: I1125 13:57:22.280486 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97dcf135-5184-42db-b23e-b39dcbe81c99-kube-api-access-j2pzt" (OuterVolumeSpecName: "kube-api-access-j2pzt") pod "97dcf135-5184-42db-b23e-b39dcbe81c99" (UID: "97dcf135-5184-42db-b23e-b39dcbe81c99"). InnerVolumeSpecName "kube-api-access-j2pzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:57:22 crc kubenswrapper[4702]: I1125 13:57:22.291977 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97dcf135-5184-42db-b23e-b39dcbe81c99-util" (OuterVolumeSpecName: "util") pod "97dcf135-5184-42db-b23e-b39dcbe81c99" (UID: "97dcf135-5184-42db-b23e-b39dcbe81c99"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:57:22 crc kubenswrapper[4702]: I1125 13:57:22.378664 4702 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/97dcf135-5184-42db-b23e-b39dcbe81c99-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 13:57:22 crc kubenswrapper[4702]: I1125 13:57:22.378703 4702 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/97dcf135-5184-42db-b23e-b39dcbe81c99-util\") on node \"crc\" DevicePath \"\"" Nov 25 13:57:22 crc kubenswrapper[4702]: I1125 13:57:22.378714 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2pzt\" (UniqueName: \"kubernetes.io/projected/97dcf135-5184-42db-b23e-b39dcbe81c99-kube-api-access-j2pzt\") on node \"crc\" DevicePath \"\"" Nov 25 13:57:22 crc kubenswrapper[4702]: I1125 13:57:22.976825 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7" event={"ID":"97dcf135-5184-42db-b23e-b39dcbe81c99","Type":"ContainerDied","Data":"b1b45405c2ff33900b91520d291888a0823b1a3412db8369dbb0bb37b166cde0"} Nov 25 13:57:22 crc kubenswrapper[4702]: I1125 13:57:22.976865 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1b45405c2ff33900b91520d291888a0823b1a3412db8369dbb0bb37b166cde0" Nov 25 13:57:22 crc kubenswrapper[4702]: I1125 13:57:22.977179 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7" Nov 25 13:57:30 crc kubenswrapper[4702]: I1125 13:57:30.490317 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-7f98f4d986-wh2hc"] Nov 25 13:57:30 crc kubenswrapper[4702]: E1125 13:57:30.491501 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97dcf135-5184-42db-b23e-b39dcbe81c99" containerName="util" Nov 25 13:57:30 crc kubenswrapper[4702]: I1125 13:57:30.491519 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="97dcf135-5184-42db-b23e-b39dcbe81c99" containerName="util" Nov 25 13:57:30 crc kubenswrapper[4702]: E1125 13:57:30.491543 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97dcf135-5184-42db-b23e-b39dcbe81c99" containerName="pull" Nov 25 13:57:30 crc kubenswrapper[4702]: I1125 13:57:30.491550 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="97dcf135-5184-42db-b23e-b39dcbe81c99" containerName="pull" Nov 25 13:57:30 crc kubenswrapper[4702]: E1125 13:57:30.491560 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97dcf135-5184-42db-b23e-b39dcbe81c99" containerName="extract" Nov 25 13:57:30 crc kubenswrapper[4702]: I1125 13:57:30.491568 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="97dcf135-5184-42db-b23e-b39dcbe81c99" containerName="extract" Nov 25 13:57:30 crc kubenswrapper[4702]: I1125 13:57:30.491687 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="97dcf135-5184-42db-b23e-b39dcbe81c99" containerName="extract" Nov 25 13:57:30 crc kubenswrapper[4702]: I1125 13:57:30.492312 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7f98f4d986-wh2hc" Nov 25 13:57:30 crc kubenswrapper[4702]: I1125 13:57:30.504822 4702 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 25 13:57:30 crc kubenswrapper[4702]: I1125 13:57:30.504830 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 25 13:57:30 crc kubenswrapper[4702]: I1125 13:57:30.505068 4702 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 25 13:57:30 crc kubenswrapper[4702]: I1125 13:57:30.505178 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 25 13:57:30 crc kubenswrapper[4702]: I1125 13:57:30.505244 4702 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-8kkdq" Nov 25 13:57:30 crc kubenswrapper[4702]: I1125 13:57:30.516977 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7f98f4d986-wh2hc"] Nov 25 13:57:30 crc kubenswrapper[4702]: I1125 13:57:30.593781 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sh6wl\" (UniqueName: \"kubernetes.io/projected/422ee193-46e3-47d9-8452-29cc5a0fa7c8-kube-api-access-sh6wl\") pod \"metallb-operator-controller-manager-7f98f4d986-wh2hc\" (UID: \"422ee193-46e3-47d9-8452-29cc5a0fa7c8\") " pod="metallb-system/metallb-operator-controller-manager-7f98f4d986-wh2hc" Nov 25 13:57:30 crc kubenswrapper[4702]: I1125 13:57:30.593920 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/422ee193-46e3-47d9-8452-29cc5a0fa7c8-apiservice-cert\") pod \"metallb-operator-controller-manager-7f98f4d986-wh2hc\" (UID: \"422ee193-46e3-47d9-8452-29cc5a0fa7c8\") " pod="metallb-system/metallb-operator-controller-manager-7f98f4d986-wh2hc" Nov 25 13:57:30 crc kubenswrapper[4702]: I1125 13:57:30.593960 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/422ee193-46e3-47d9-8452-29cc5a0fa7c8-webhook-cert\") pod \"metallb-operator-controller-manager-7f98f4d986-wh2hc\" (UID: \"422ee193-46e3-47d9-8452-29cc5a0fa7c8\") " pod="metallb-system/metallb-operator-controller-manager-7f98f4d986-wh2hc" Nov 25 13:57:30 crc kubenswrapper[4702]: I1125 13:57:30.695539 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/422ee193-46e3-47d9-8452-29cc5a0fa7c8-apiservice-cert\") pod \"metallb-operator-controller-manager-7f98f4d986-wh2hc\" (UID: \"422ee193-46e3-47d9-8452-29cc5a0fa7c8\") " pod="metallb-system/metallb-operator-controller-manager-7f98f4d986-wh2hc" Nov 25 13:57:30 crc kubenswrapper[4702]: I1125 13:57:30.695624 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/422ee193-46e3-47d9-8452-29cc5a0fa7c8-webhook-cert\") pod \"metallb-operator-controller-manager-7f98f4d986-wh2hc\" (UID: \"422ee193-46e3-47d9-8452-29cc5a0fa7c8\") " pod="metallb-system/metallb-operator-controller-manager-7f98f4d986-wh2hc" Nov 25 13:57:30 crc kubenswrapper[4702]: I1125 13:57:30.695697 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sh6wl\" (UniqueName: \"kubernetes.io/projected/422ee193-46e3-47d9-8452-29cc5a0fa7c8-kube-api-access-sh6wl\") pod \"metallb-operator-controller-manager-7f98f4d986-wh2hc\" (UID: \"422ee193-46e3-47d9-8452-29cc5a0fa7c8\") " pod="metallb-system/metallb-operator-controller-manager-7f98f4d986-wh2hc" Nov 25 13:57:30 crc kubenswrapper[4702]: I1125 13:57:30.703036 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/422ee193-46e3-47d9-8452-29cc5a0fa7c8-apiservice-cert\") pod \"metallb-operator-controller-manager-7f98f4d986-wh2hc\" (UID: \"422ee193-46e3-47d9-8452-29cc5a0fa7c8\") " pod="metallb-system/metallb-operator-controller-manager-7f98f4d986-wh2hc" Nov 25 13:57:30 crc kubenswrapper[4702]: I1125 13:57:30.704594 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/422ee193-46e3-47d9-8452-29cc5a0fa7c8-webhook-cert\") pod \"metallb-operator-controller-manager-7f98f4d986-wh2hc\" (UID: \"422ee193-46e3-47d9-8452-29cc5a0fa7c8\") " pod="metallb-system/metallb-operator-controller-manager-7f98f4d986-wh2hc" Nov 25 13:57:30 crc kubenswrapper[4702]: I1125 13:57:30.718779 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sh6wl\" (UniqueName: \"kubernetes.io/projected/422ee193-46e3-47d9-8452-29cc5a0fa7c8-kube-api-access-sh6wl\") pod \"metallb-operator-controller-manager-7f98f4d986-wh2hc\" (UID: \"422ee193-46e3-47d9-8452-29cc5a0fa7c8\") " pod="metallb-system/metallb-operator-controller-manager-7f98f4d986-wh2hc" Nov 25 13:57:30 crc kubenswrapper[4702]: I1125 13:57:30.744586 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-757b88d98b-pkltf"] Nov 25 13:57:30 crc kubenswrapper[4702]: I1125 13:57:30.745200 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-757b88d98b-pkltf" Nov 25 13:57:30 crc kubenswrapper[4702]: I1125 13:57:30.748026 4702 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 25 13:57:30 crc kubenswrapper[4702]: I1125 13:57:30.748041 4702 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 25 13:57:30 crc kubenswrapper[4702]: I1125 13:57:30.751303 4702 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-9mhjd" Nov 25 13:57:30 crc kubenswrapper[4702]: I1125 13:57:30.768868 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-757b88d98b-pkltf"] Nov 25 13:57:30 crc kubenswrapper[4702]: I1125 13:57:30.817248 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7f98f4d986-wh2hc" Nov 25 13:57:30 crc kubenswrapper[4702]: I1125 13:57:30.897957 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/28d9e60e-bc7b-4268-b855-93c5c98a19bb-webhook-cert\") pod \"metallb-operator-webhook-server-757b88d98b-pkltf\" (UID: \"28d9e60e-bc7b-4268-b855-93c5c98a19bb\") " pod="metallb-system/metallb-operator-webhook-server-757b88d98b-pkltf" Nov 25 13:57:30 crc kubenswrapper[4702]: I1125 13:57:30.898034 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/28d9e60e-bc7b-4268-b855-93c5c98a19bb-apiservice-cert\") pod \"metallb-operator-webhook-server-757b88d98b-pkltf\" (UID: \"28d9e60e-bc7b-4268-b855-93c5c98a19bb\") " pod="metallb-system/metallb-operator-webhook-server-757b88d98b-pkltf" Nov 25 13:57:30 crc kubenswrapper[4702]: I1125 13:57:30.898077 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8slzc\" (UniqueName: \"kubernetes.io/projected/28d9e60e-bc7b-4268-b855-93c5c98a19bb-kube-api-access-8slzc\") pod \"metallb-operator-webhook-server-757b88d98b-pkltf\" (UID: \"28d9e60e-bc7b-4268-b855-93c5c98a19bb\") " pod="metallb-system/metallb-operator-webhook-server-757b88d98b-pkltf" Nov 25 13:57:31 crc kubenswrapper[4702]: I1125 13:57:30.999893 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8slzc\" (UniqueName: \"kubernetes.io/projected/28d9e60e-bc7b-4268-b855-93c5c98a19bb-kube-api-access-8slzc\") pod \"metallb-operator-webhook-server-757b88d98b-pkltf\" (UID: \"28d9e60e-bc7b-4268-b855-93c5c98a19bb\") " pod="metallb-system/metallb-operator-webhook-server-757b88d98b-pkltf" Nov 25 13:57:31 crc kubenswrapper[4702]: I1125 13:57:31.000480 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/28d9e60e-bc7b-4268-b855-93c5c98a19bb-webhook-cert\") pod \"metallb-operator-webhook-server-757b88d98b-pkltf\" (UID: \"28d9e60e-bc7b-4268-b855-93c5c98a19bb\") " pod="metallb-system/metallb-operator-webhook-server-757b88d98b-pkltf" Nov 25 13:57:31 crc kubenswrapper[4702]: I1125 13:57:31.000543 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/28d9e60e-bc7b-4268-b855-93c5c98a19bb-apiservice-cert\") pod \"metallb-operator-webhook-server-757b88d98b-pkltf\" (UID: \"28d9e60e-bc7b-4268-b855-93c5c98a19bb\") " pod="metallb-system/metallb-operator-webhook-server-757b88d98b-pkltf" Nov 25 13:57:31 crc kubenswrapper[4702]: I1125 13:57:31.010386 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/28d9e60e-bc7b-4268-b855-93c5c98a19bb-apiservice-cert\") pod \"metallb-operator-webhook-server-757b88d98b-pkltf\" (UID: \"28d9e60e-bc7b-4268-b855-93c5c98a19bb\") " pod="metallb-system/metallb-operator-webhook-server-757b88d98b-pkltf" Nov 25 13:57:31 crc kubenswrapper[4702]: I1125 13:57:31.013016 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/28d9e60e-bc7b-4268-b855-93c5c98a19bb-webhook-cert\") pod \"metallb-operator-webhook-server-757b88d98b-pkltf\" (UID: \"28d9e60e-bc7b-4268-b855-93c5c98a19bb\") " pod="metallb-system/metallb-operator-webhook-server-757b88d98b-pkltf" Nov 25 13:57:31 crc kubenswrapper[4702]: I1125 13:57:31.036985 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8slzc\" (UniqueName: \"kubernetes.io/projected/28d9e60e-bc7b-4268-b855-93c5c98a19bb-kube-api-access-8slzc\") pod \"metallb-operator-webhook-server-757b88d98b-pkltf\" (UID: \"28d9e60e-bc7b-4268-b855-93c5c98a19bb\") " pod="metallb-system/metallb-operator-webhook-server-757b88d98b-pkltf" Nov 25 13:57:31 crc kubenswrapper[4702]: I1125 13:57:31.068557 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-757b88d98b-pkltf" Nov 25 13:57:31 crc kubenswrapper[4702]: I1125 13:57:31.159516 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7f98f4d986-wh2hc"] Nov 25 13:57:31 crc kubenswrapper[4702]: W1125 13:57:31.169645 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod422ee193_46e3_47d9_8452_29cc5a0fa7c8.slice/crio-b74565d60683c47320931b64e10a4ece82a0a6791aba22564d349d5006bd1550 WatchSource:0}: Error finding container b74565d60683c47320931b64e10a4ece82a0a6791aba22564d349d5006bd1550: Status 404 returned error can't find the container with id b74565d60683c47320931b64e10a4ece82a0a6791aba22564d349d5006bd1550 Nov 25 13:57:31 crc kubenswrapper[4702]: I1125 13:57:31.378975 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-757b88d98b-pkltf"] Nov 25 13:57:31 crc kubenswrapper[4702]: W1125 13:57:31.387912 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28d9e60e_bc7b_4268_b855_93c5c98a19bb.slice/crio-ba50e44cfe9aa3355af69edbf31e24f42657b1bd054aa3ad1fad038a8c9bfe48 WatchSource:0}: Error finding container ba50e44cfe9aa3355af69edbf31e24f42657b1bd054aa3ad1fad038a8c9bfe48: Status 404 returned error can't find the container with id ba50e44cfe9aa3355af69edbf31e24f42657b1bd054aa3ad1fad038a8c9bfe48 Nov 25 13:57:32 crc kubenswrapper[4702]: I1125 13:57:32.022339 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-757b88d98b-pkltf" event={"ID":"28d9e60e-bc7b-4268-b855-93c5c98a19bb","Type":"ContainerStarted","Data":"ba50e44cfe9aa3355af69edbf31e24f42657b1bd054aa3ad1fad038a8c9bfe48"} Nov 25 13:57:32 crc kubenswrapper[4702]: I1125 13:57:32.025630 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7f98f4d986-wh2hc" event={"ID":"422ee193-46e3-47d9-8452-29cc5a0fa7c8","Type":"ContainerStarted","Data":"b74565d60683c47320931b64e10a4ece82a0a6791aba22564d349d5006bd1550"} Nov 25 13:57:36 crc kubenswrapper[4702]: I1125 13:57:36.053649 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7f98f4d986-wh2hc" event={"ID":"422ee193-46e3-47d9-8452-29cc5a0fa7c8","Type":"ContainerStarted","Data":"e6bea9cfccbca48b2921761ea9c0a00d6b73644a0478ad6d1658ad994d6c6258"} Nov 25 13:57:36 crc kubenswrapper[4702]: I1125 13:57:36.054249 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-7f98f4d986-wh2hc" Nov 25 13:57:36 crc kubenswrapper[4702]: I1125 13:57:36.077886 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-7f98f4d986-wh2hc" podStartSLOduration=1.368448386 podStartE2EDuration="6.07786807s" podCreationTimestamp="2025-11-25 13:57:30 +0000 UTC" firstStartedPulling="2025-11-25 13:57:31.174125776 +0000 UTC m=+660.336776226" lastFinishedPulling="2025-11-25 13:57:35.88354546 +0000 UTC m=+665.046195910" observedRunningTime="2025-11-25 13:57:36.074163203 +0000 UTC m=+665.236813653" watchObservedRunningTime="2025-11-25 13:57:36.07786807 +0000 UTC m=+665.240518520" Nov 25 13:57:37 crc kubenswrapper[4702]: I1125 13:57:37.060012 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-757b88d98b-pkltf" event={"ID":"28d9e60e-bc7b-4268-b855-93c5c98a19bb","Type":"ContainerStarted","Data":"c84c92c7e7e134dd6df6f3d3e1d966818581388d67a39a02218b9fa1669a2148"} Nov 25 13:57:37 crc kubenswrapper[4702]: I1125 13:57:37.081962 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-757b88d98b-pkltf" podStartSLOduration=2.063918324 podStartE2EDuration="7.081945486s" podCreationTimestamp="2025-11-25 13:57:30 +0000 UTC" firstStartedPulling="2025-11-25 13:57:31.391278846 +0000 UTC m=+660.553929296" lastFinishedPulling="2025-11-25 13:57:36.409306008 +0000 UTC m=+665.571956458" observedRunningTime="2025-11-25 13:57:37.078265231 +0000 UTC m=+666.240915681" watchObservedRunningTime="2025-11-25 13:57:37.081945486 +0000 UTC m=+666.244595936" Nov 25 13:57:38 crc kubenswrapper[4702]: I1125 13:57:38.064209 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-757b88d98b-pkltf" Nov 25 13:57:51 crc kubenswrapper[4702]: I1125 13:57:51.080615 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-757b88d98b-pkltf" Nov 25 13:58:10 crc kubenswrapper[4702]: I1125 13:58:10.820489 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-7f98f4d986-wh2hc" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.553113 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-hpc6n"] Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.556102 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-hpc6n" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.561155 4702 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.561530 4702 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-hblfp" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.562347 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.565992 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-btnfr"] Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.567009 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-btnfr" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.569579 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3-metrics\") pod \"frr-k8s-hpc6n\" (UID: \"5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3\") " pod="metallb-system/frr-k8s-hpc6n" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.569646 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3-frr-startup\") pod \"frr-k8s-hpc6n\" (UID: \"5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3\") " pod="metallb-system/frr-k8s-hpc6n" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.569683 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3-frr-conf\") pod \"frr-k8s-hpc6n\" (UID: \"5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3\") " pod="metallb-system/frr-k8s-hpc6n" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.569717 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3-frr-sockets\") pod \"frr-k8s-hpc6n\" (UID: \"5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3\") " pod="metallb-system/frr-k8s-hpc6n" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.569749 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c0aec7c5-7588-4f42-a64a-9bbe4c07ea1c-cert\") pod \"frr-k8s-webhook-server-6998585d5-btnfr\" (UID: \"c0aec7c5-7588-4f42-a64a-9bbe4c07ea1c\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-btnfr" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.569781 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62xv9\" (UniqueName: \"kubernetes.io/projected/c0aec7c5-7588-4f42-a64a-9bbe4c07ea1c-kube-api-access-62xv9\") pod \"frr-k8s-webhook-server-6998585d5-btnfr\" (UID: \"c0aec7c5-7588-4f42-a64a-9bbe4c07ea1c\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-btnfr" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.569812 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3-reloader\") pod \"frr-k8s-hpc6n\" (UID: \"5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3\") " pod="metallb-system/frr-k8s-hpc6n" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.569842 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qqwt\" (UniqueName: \"kubernetes.io/projected/5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3-kube-api-access-9qqwt\") pod \"frr-k8s-hpc6n\" (UID: \"5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3\") " pod="metallb-system/frr-k8s-hpc6n" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.569905 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3-metrics-certs\") pod \"frr-k8s-hpc6n\" (UID: \"5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3\") " pod="metallb-system/frr-k8s-hpc6n" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.572325 4702 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.583843 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-btnfr"] Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.670547 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c0aec7c5-7588-4f42-a64a-9bbe4c07ea1c-cert\") pod \"frr-k8s-webhook-server-6998585d5-btnfr\" (UID: \"c0aec7c5-7588-4f42-a64a-9bbe4c07ea1c\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-btnfr" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.670608 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62xv9\" (UniqueName: \"kubernetes.io/projected/c0aec7c5-7588-4f42-a64a-9bbe4c07ea1c-kube-api-access-62xv9\") pod \"frr-k8s-webhook-server-6998585d5-btnfr\" (UID: \"c0aec7c5-7588-4f42-a64a-9bbe4c07ea1c\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-btnfr" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.670638 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3-reloader\") pod \"frr-k8s-hpc6n\" (UID: \"5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3\") " pod="metallb-system/frr-k8s-hpc6n" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.670670 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qqwt\" (UniqueName: \"kubernetes.io/projected/5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3-kube-api-access-9qqwt\") pod \"frr-k8s-hpc6n\" (UID: \"5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3\") " pod="metallb-system/frr-k8s-hpc6n" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.670699 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3-metrics-certs\") pod \"frr-k8s-hpc6n\" (UID: \"5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3\") " pod="metallb-system/frr-k8s-hpc6n" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.670743 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3-metrics\") pod \"frr-k8s-hpc6n\" (UID: \"5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3\") " pod="metallb-system/frr-k8s-hpc6n" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.670769 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3-frr-startup\") pod \"frr-k8s-hpc6n\" (UID: \"5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3\") " pod="metallb-system/frr-k8s-hpc6n" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.670797 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3-frr-conf\") pod \"frr-k8s-hpc6n\" (UID: \"5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3\") " pod="metallb-system/frr-k8s-hpc6n" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.670827 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3-frr-sockets\") pod \"frr-k8s-hpc6n\" (UID: \"5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3\") " pod="metallb-system/frr-k8s-hpc6n" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.671328 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3-frr-sockets\") pod \"frr-k8s-hpc6n\" (UID: \"5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3\") " pod="metallb-system/frr-k8s-hpc6n" Nov 25 13:58:11 crc kubenswrapper[4702]: E1125 13:58:11.671422 4702 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Nov 25 13:58:11 crc kubenswrapper[4702]: E1125 13:58:11.671477 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3-metrics-certs podName:5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3 nodeName:}" failed. No retries permitted until 2025-11-25 13:58:12.171459304 +0000 UTC m=+701.334109754 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3-metrics-certs") pod "frr-k8s-hpc6n" (UID: "5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3") : secret "frr-k8s-certs-secret" not found Nov 25 13:58:11 crc kubenswrapper[4702]: E1125 13:58:11.671568 4702 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Nov 25 13:58:11 crc kubenswrapper[4702]: E1125 13:58:11.671675 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c0aec7c5-7588-4f42-a64a-9bbe4c07ea1c-cert podName:c0aec7c5-7588-4f42-a64a-9bbe4c07ea1c nodeName:}" failed. No retries permitted until 2025-11-25 13:58:12.171648479 +0000 UTC m=+701.334298989 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c0aec7c5-7588-4f42-a64a-9bbe4c07ea1c-cert") pod "frr-k8s-webhook-server-6998585d5-btnfr" (UID: "c0aec7c5-7588-4f42-a64a-9bbe4c07ea1c") : secret "frr-k8s-webhook-server-cert" not found Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.671851 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3-reloader\") pod \"frr-k8s-hpc6n\" (UID: \"5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3\") " pod="metallb-system/frr-k8s-hpc6n" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.672249 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3-frr-conf\") pod \"frr-k8s-hpc6n\" (UID: \"5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3\") " pod="metallb-system/frr-k8s-hpc6n" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.672344 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3-metrics\") pod \"frr-k8s-hpc6n\" (UID: \"5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3\") " pod="metallb-system/frr-k8s-hpc6n" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.673000 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3-frr-startup\") pod \"frr-k8s-hpc6n\" (UID: \"5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3\") " pod="metallb-system/frr-k8s-hpc6n" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.688019 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-45qwr"] Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.689064 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-45qwr" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.690637 4702 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-bpkd6" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.691134 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.691354 4702 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.691523 4702 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.710306 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qqwt\" (UniqueName: \"kubernetes.io/projected/5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3-kube-api-access-9qqwt\") pod \"frr-k8s-hpc6n\" (UID: \"5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3\") " pod="metallb-system/frr-k8s-hpc6n" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.713407 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-chvrp"] Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.714098 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62xv9\" (UniqueName: \"kubernetes.io/projected/c0aec7c5-7588-4f42-a64a-9bbe4c07ea1c-kube-api-access-62xv9\") pod \"frr-k8s-webhook-server-6998585d5-btnfr\" (UID: \"c0aec7c5-7588-4f42-a64a-9bbe4c07ea1c\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-btnfr" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.714214 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-chvrp" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.717461 4702 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.745952 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-chvrp"] Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.873620 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d819dd6c-359a-4449-be38-1f0b18ae64d2-cert\") pod \"controller-6c7b4b5f48-chvrp\" (UID: \"d819dd6c-359a-4449-be38-1f0b18ae64d2\") " pod="metallb-system/controller-6c7b4b5f48-chvrp" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.873727 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/4f16595d-125a-4e11-a745-abea073fe836-metallb-excludel2\") pod \"speaker-45qwr\" (UID: \"4f16595d-125a-4e11-a745-abea073fe836\") " pod="metallb-system/speaker-45qwr" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.873774 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d819dd6c-359a-4449-be38-1f0b18ae64d2-metrics-certs\") pod \"controller-6c7b4b5f48-chvrp\" (UID: \"d819dd6c-359a-4449-be38-1f0b18ae64d2\") " pod="metallb-system/controller-6c7b4b5f48-chvrp" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.874104 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/4f16595d-125a-4e11-a745-abea073fe836-memberlist\") pod \"speaker-45qwr\" (UID: \"4f16595d-125a-4e11-a745-abea073fe836\") " pod="metallb-system/speaker-45qwr" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.874451 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76vvv\" (UniqueName: \"kubernetes.io/projected/4f16595d-125a-4e11-a745-abea073fe836-kube-api-access-76vvv\") pod \"speaker-45qwr\" (UID: \"4f16595d-125a-4e11-a745-abea073fe836\") " pod="metallb-system/speaker-45qwr" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.874514 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ss6l\" (UniqueName: \"kubernetes.io/projected/d819dd6c-359a-4449-be38-1f0b18ae64d2-kube-api-access-8ss6l\") pod \"controller-6c7b4b5f48-chvrp\" (UID: \"d819dd6c-359a-4449-be38-1f0b18ae64d2\") " pod="metallb-system/controller-6c7b4b5f48-chvrp" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.874532 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4f16595d-125a-4e11-a745-abea073fe836-metrics-certs\") pod \"speaker-45qwr\" (UID: \"4f16595d-125a-4e11-a745-abea073fe836\") " pod="metallb-system/speaker-45qwr" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.975585 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d819dd6c-359a-4449-be38-1f0b18ae64d2-metrics-certs\") pod \"controller-6c7b4b5f48-chvrp\" (UID: \"d819dd6c-359a-4449-be38-1f0b18ae64d2\") " pod="metallb-system/controller-6c7b4b5f48-chvrp" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.975691 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/4f16595d-125a-4e11-a745-abea073fe836-memberlist\") pod \"speaker-45qwr\" (UID: \"4f16595d-125a-4e11-a745-abea073fe836\") " pod="metallb-system/speaker-45qwr" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.975720 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76vvv\" (UniqueName: \"kubernetes.io/projected/4f16595d-125a-4e11-a745-abea073fe836-kube-api-access-76vvv\") pod \"speaker-45qwr\" (UID: \"4f16595d-125a-4e11-a745-abea073fe836\") " pod="metallb-system/speaker-45qwr" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.975750 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ss6l\" (UniqueName: \"kubernetes.io/projected/d819dd6c-359a-4449-be38-1f0b18ae64d2-kube-api-access-8ss6l\") pod \"controller-6c7b4b5f48-chvrp\" (UID: \"d819dd6c-359a-4449-be38-1f0b18ae64d2\") " pod="metallb-system/controller-6c7b4b5f48-chvrp" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.975781 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4f16595d-125a-4e11-a745-abea073fe836-metrics-certs\") pod \"speaker-45qwr\" (UID: \"4f16595d-125a-4e11-a745-abea073fe836\") " pod="metallb-system/speaker-45qwr" Nov 25 13:58:11 crc kubenswrapper[4702]: E1125 13:58:11.975866 4702 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 25 13:58:11 crc kubenswrapper[4702]: E1125 13:58:11.975936 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4f16595d-125a-4e11-a745-abea073fe836-memberlist podName:4f16595d-125a-4e11-a745-abea073fe836 nodeName:}" failed. No retries permitted until 2025-11-25 13:58:12.475917044 +0000 UTC m=+701.638567504 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/4f16595d-125a-4e11-a745-abea073fe836-memberlist") pod "speaker-45qwr" (UID: "4f16595d-125a-4e11-a745-abea073fe836") : secret "metallb-memberlist" not found Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.976563 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d819dd6c-359a-4449-be38-1f0b18ae64d2-cert\") pod \"controller-6c7b4b5f48-chvrp\" (UID: \"d819dd6c-359a-4449-be38-1f0b18ae64d2\") " pod="metallb-system/controller-6c7b4b5f48-chvrp" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.976655 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/4f16595d-125a-4e11-a745-abea073fe836-metallb-excludel2\") pod \"speaker-45qwr\" (UID: \"4f16595d-125a-4e11-a745-abea073fe836\") " pod="metallb-system/speaker-45qwr" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.977609 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/4f16595d-125a-4e11-a745-abea073fe836-metallb-excludel2\") pod \"speaker-45qwr\" (UID: \"4f16595d-125a-4e11-a745-abea073fe836\") " pod="metallb-system/speaker-45qwr" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.979737 4702 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 25 13:58:11 crc kubenswrapper[4702]: I1125 13:58:11.990521 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d819dd6c-359a-4449-be38-1f0b18ae64d2-cert\") pod \"controller-6c7b4b5f48-chvrp\" (UID: \"d819dd6c-359a-4449-be38-1f0b18ae64d2\") " pod="metallb-system/controller-6c7b4b5f48-chvrp" Nov 25 13:58:12 crc kubenswrapper[4702]: I1125 13:58:12.018868 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4f16595d-125a-4e11-a745-abea073fe836-metrics-certs\") pod \"speaker-45qwr\" (UID: \"4f16595d-125a-4e11-a745-abea073fe836\") " pod="metallb-system/speaker-45qwr" Nov 25 13:58:12 crc kubenswrapper[4702]: I1125 13:58:12.018992 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d819dd6c-359a-4449-be38-1f0b18ae64d2-metrics-certs\") pod \"controller-6c7b4b5f48-chvrp\" (UID: \"d819dd6c-359a-4449-be38-1f0b18ae64d2\") " pod="metallb-system/controller-6c7b4b5f48-chvrp" Nov 25 13:58:12 crc kubenswrapper[4702]: I1125 13:58:12.022146 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ss6l\" (UniqueName: \"kubernetes.io/projected/d819dd6c-359a-4449-be38-1f0b18ae64d2-kube-api-access-8ss6l\") pod \"controller-6c7b4b5f48-chvrp\" (UID: \"d819dd6c-359a-4449-be38-1f0b18ae64d2\") " pod="metallb-system/controller-6c7b4b5f48-chvrp" Nov 25 13:58:12 crc kubenswrapper[4702]: I1125 13:58:12.024074 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76vvv\" (UniqueName: \"kubernetes.io/projected/4f16595d-125a-4e11-a745-abea073fe836-kube-api-access-76vvv\") pod \"speaker-45qwr\" (UID: \"4f16595d-125a-4e11-a745-abea073fe836\") " pod="metallb-system/speaker-45qwr" Nov 25 13:58:12 crc kubenswrapper[4702]: I1125 13:58:12.077640 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-chvrp" Nov 25 13:58:12 crc kubenswrapper[4702]: I1125 13:58:12.182086 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c0aec7c5-7588-4f42-a64a-9bbe4c07ea1c-cert\") pod \"frr-k8s-webhook-server-6998585d5-btnfr\" (UID: \"c0aec7c5-7588-4f42-a64a-9bbe4c07ea1c\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-btnfr" Nov 25 13:58:12 crc kubenswrapper[4702]: I1125 13:58:12.182373 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3-metrics-certs\") pod \"frr-k8s-hpc6n\" (UID: \"5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3\") " pod="metallb-system/frr-k8s-hpc6n" Nov 25 13:58:12 crc kubenswrapper[4702]: I1125 13:58:12.187672 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3-metrics-certs\") pod \"frr-k8s-hpc6n\" (UID: \"5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3\") " pod="metallb-system/frr-k8s-hpc6n" Nov 25 13:58:12 crc kubenswrapper[4702]: I1125 13:58:12.192019 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-hpc6n" Nov 25 13:58:12 crc kubenswrapper[4702]: I1125 13:58:12.193783 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c0aec7c5-7588-4f42-a64a-9bbe4c07ea1c-cert\") pod \"frr-k8s-webhook-server-6998585d5-btnfr\" (UID: \"c0aec7c5-7588-4f42-a64a-9bbe4c07ea1c\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-btnfr" Nov 25 13:58:12 crc kubenswrapper[4702]: I1125 13:58:12.205813 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-btnfr" Nov 25 13:58:12 crc kubenswrapper[4702]: I1125 13:58:12.307820 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-chvrp"] Nov 25 13:58:12 crc kubenswrapper[4702]: W1125 13:58:12.315918 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd819dd6c_359a_4449_be38_1f0b18ae64d2.slice/crio-7d95c9bb5ef93fbfca1425d281e6e1d145c711e6c4a150b32015f7694669b83b WatchSource:0}: Error finding container 7d95c9bb5ef93fbfca1425d281e6e1d145c711e6c4a150b32015f7694669b83b: Status 404 returned error can't find the container with id 7d95c9bb5ef93fbfca1425d281e6e1d145c711e6c4a150b32015f7694669b83b Nov 25 13:58:12 crc kubenswrapper[4702]: I1125 13:58:12.490676 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/4f16595d-125a-4e11-a745-abea073fe836-memberlist\") pod \"speaker-45qwr\" (UID: \"4f16595d-125a-4e11-a745-abea073fe836\") " pod="metallb-system/speaker-45qwr" Nov 25 13:58:12 crc kubenswrapper[4702]: E1125 13:58:12.490887 4702 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 25 13:58:12 crc kubenswrapper[4702]: E1125 13:58:12.490975 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4f16595d-125a-4e11-a745-abea073fe836-memberlist podName:4f16595d-125a-4e11-a745-abea073fe836 nodeName:}" failed. No retries permitted until 2025-11-25 13:58:13.490954034 +0000 UTC m=+702.653604484 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/4f16595d-125a-4e11-a745-abea073fe836-memberlist") pod "speaker-45qwr" (UID: "4f16595d-125a-4e11-a745-abea073fe836") : secret "metallb-memberlist" not found Nov 25 13:58:12 crc kubenswrapper[4702]: I1125 13:58:12.491207 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-btnfr"] Nov 25 13:58:12 crc kubenswrapper[4702]: W1125 13:58:12.495079 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0aec7c5_7588_4f42_a64a_9bbe4c07ea1c.slice/crio-05588613b5b365e8535074ffd5147da817baa090aee1caf38b80e17372ccd0fb WatchSource:0}: Error finding container 05588613b5b365e8535074ffd5147da817baa090aee1caf38b80e17372ccd0fb: Status 404 returned error can't find the container with id 05588613b5b365e8535074ffd5147da817baa090aee1caf38b80e17372ccd0fb Nov 25 13:58:13 crc kubenswrapper[4702]: I1125 13:58:13.301339 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-btnfr" event={"ID":"c0aec7c5-7588-4f42-a64a-9bbe4c07ea1c","Type":"ContainerStarted","Data":"05588613b5b365e8535074ffd5147da817baa090aee1caf38b80e17372ccd0fb"} Nov 25 13:58:13 crc kubenswrapper[4702]: I1125 13:58:13.302827 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hpc6n" event={"ID":"5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3","Type":"ContainerStarted","Data":"65b0282afe2cece8a2165704b0dd05cd5addbccab509b0b727ef282389298202"} Nov 25 13:58:13 crc kubenswrapper[4702]: I1125 13:58:13.305123 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-chvrp" event={"ID":"d819dd6c-359a-4449-be38-1f0b18ae64d2","Type":"ContainerStarted","Data":"7803331f9d8b050c74a5e544cbf4a3d4cfca6c555cff6d2ed9b25e01cbea66f1"} Nov 25 13:58:13 crc kubenswrapper[4702]: I1125 13:58:13.305186 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-chvrp" event={"ID":"d819dd6c-359a-4449-be38-1f0b18ae64d2","Type":"ContainerStarted","Data":"7d95c9bb5ef93fbfca1425d281e6e1d145c711e6c4a150b32015f7694669b83b"} Nov 25 13:58:13 crc kubenswrapper[4702]: I1125 13:58:13.504030 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/4f16595d-125a-4e11-a745-abea073fe836-memberlist\") pod \"speaker-45qwr\" (UID: \"4f16595d-125a-4e11-a745-abea073fe836\") " pod="metallb-system/speaker-45qwr" Nov 25 13:58:13 crc kubenswrapper[4702]: I1125 13:58:13.512195 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/4f16595d-125a-4e11-a745-abea073fe836-memberlist\") pod \"speaker-45qwr\" (UID: \"4f16595d-125a-4e11-a745-abea073fe836\") " pod="metallb-system/speaker-45qwr" Nov 25 13:58:13 crc kubenswrapper[4702]: I1125 13:58:13.567124 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-45qwr" Nov 25 13:58:14 crc kubenswrapper[4702]: I1125 13:58:14.313339 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-45qwr" event={"ID":"4f16595d-125a-4e11-a745-abea073fe836","Type":"ContainerStarted","Data":"45e14086b56dd0b76bb6eedf65fcf807ee67cfb1d7fcd1570974035afc5907c6"} Nov 25 13:58:14 crc kubenswrapper[4702]: I1125 13:58:14.313772 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-45qwr" event={"ID":"4f16595d-125a-4e11-a745-abea073fe836","Type":"ContainerStarted","Data":"98cef5f465fcedbb9470aef084e0783f9ec1b3f3105290d04ab281ea69a68679"} Nov 25 13:58:17 crc kubenswrapper[4702]: I1125 13:58:17.342511 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-45qwr" event={"ID":"4f16595d-125a-4e11-a745-abea073fe836","Type":"ContainerStarted","Data":"6e5159fbc665bd7e9ed763bab23a75a1357ef614fed115f8060f90cd3429920b"} Nov 25 13:58:17 crc kubenswrapper[4702]: I1125 13:58:17.342692 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-45qwr" Nov 25 13:58:17 crc kubenswrapper[4702]: I1125 13:58:17.345608 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-chvrp" event={"ID":"d819dd6c-359a-4449-be38-1f0b18ae64d2","Type":"ContainerStarted","Data":"399e6d4e6029950a1d6a73da5957df195ee0e54db31ee308116d71ebe4bbde44"} Nov 25 13:58:17 crc kubenswrapper[4702]: I1125 13:58:17.346321 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-chvrp" Nov 25 13:58:17 crc kubenswrapper[4702]: I1125 13:58:17.392969 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-45qwr" podStartSLOduration=4.017115004 podStartE2EDuration="6.392946853s" podCreationTimestamp="2025-11-25 13:58:11 +0000 UTC" firstStartedPulling="2025-11-25 13:58:14.053423232 +0000 UTC m=+703.216073682" lastFinishedPulling="2025-11-25 13:58:16.429255081 +0000 UTC m=+705.591905531" observedRunningTime="2025-11-25 13:58:17.360343289 +0000 UTC m=+706.522993739" watchObservedRunningTime="2025-11-25 13:58:17.392946853 +0000 UTC m=+706.555597303" Nov 25 13:58:17 crc kubenswrapper[4702]: I1125 13:58:17.395474 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-chvrp" podStartSLOduration=2.433951469 podStartE2EDuration="6.395441337s" podCreationTimestamp="2025-11-25 13:58:11 +0000 UTC" firstStartedPulling="2025-11-25 13:58:12.442733475 +0000 UTC m=+701.605383925" lastFinishedPulling="2025-11-25 13:58:16.404223353 +0000 UTC m=+705.566873793" observedRunningTime="2025-11-25 13:58:17.390485329 +0000 UTC m=+706.553135779" watchObservedRunningTime="2025-11-25 13:58:17.395441337 +0000 UTC m=+706.558091787" Nov 25 13:58:17 crc kubenswrapper[4702]: I1125 13:58:17.693622 4702 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 13:58:21 crc kubenswrapper[4702]: I1125 13:58:21.375284 4702 generic.go:334] "Generic (PLEG): container finished" podID="5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3" containerID="7db8eb4965aa3cd2adcee2dabf6030f17f907e378d3dddb1edc282137ac1acbf" exitCode=0 Nov 25 13:58:21 crc kubenswrapper[4702]: I1125 13:58:21.375359 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hpc6n" event={"ID":"5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3","Type":"ContainerDied","Data":"7db8eb4965aa3cd2adcee2dabf6030f17f907e378d3dddb1edc282137ac1acbf"} Nov 25 13:58:21 crc kubenswrapper[4702]: I1125 13:58:21.379852 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-btnfr" event={"ID":"c0aec7c5-7588-4f42-a64a-9bbe4c07ea1c","Type":"ContainerStarted","Data":"1fde441e404c82813f9ae425437b4ef79ec17f57e9f8e1f8b140224f155adb97"} Nov 25 13:58:21 crc kubenswrapper[4702]: I1125 13:58:21.380246 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-btnfr" Nov 25 13:58:21 crc kubenswrapper[4702]: I1125 13:58:21.448817 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-btnfr" podStartSLOduration=2.423123899 podStartE2EDuration="10.448791162s" podCreationTimestamp="2025-11-25 13:58:11 +0000 UTC" firstStartedPulling="2025-11-25 13:58:12.497314138 +0000 UTC m=+701.659964588" lastFinishedPulling="2025-11-25 13:58:20.522981401 +0000 UTC m=+709.685631851" observedRunningTime="2025-11-25 13:58:21.438283531 +0000 UTC m=+710.600933991" watchObservedRunningTime="2025-11-25 13:58:21.448791162 +0000 UTC m=+710.611441642" Nov 25 13:58:22 crc kubenswrapper[4702]: I1125 13:58:22.082318 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-chvrp" Nov 25 13:58:22 crc kubenswrapper[4702]: I1125 13:58:22.395581 4702 generic.go:334] "Generic (PLEG): container finished" podID="5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3" containerID="704630363aaeadf29003c02216acddec5ee66364f560c854537bddf05d7ac919" exitCode=0 Nov 25 13:58:22 crc kubenswrapper[4702]: I1125 13:58:22.395666 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hpc6n" event={"ID":"5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3","Type":"ContainerDied","Data":"704630363aaeadf29003c02216acddec5ee66364f560c854537bddf05d7ac919"} Nov 25 13:58:23 crc kubenswrapper[4702]: I1125 13:58:23.404327 4702 generic.go:334] "Generic (PLEG): container finished" podID="5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3" containerID="09b607fcdd2cd6672c27a283fb37f4c4fe913d6287acaa82727d9c5e6afb8eaa" exitCode=0 Nov 25 13:58:23 crc kubenswrapper[4702]: I1125 13:58:23.404390 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hpc6n" event={"ID":"5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3","Type":"ContainerDied","Data":"09b607fcdd2cd6672c27a283fb37f4c4fe913d6287acaa82727d9c5e6afb8eaa"} Nov 25 13:58:24 crc kubenswrapper[4702]: I1125 13:58:24.418219 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hpc6n" event={"ID":"5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3","Type":"ContainerStarted","Data":"d0caccee826d79af8cdfd6bcb1485a9ccf005c58b363dcd2be2095acadc90084"} Nov 25 13:58:24 crc kubenswrapper[4702]: I1125 13:58:24.418810 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hpc6n" event={"ID":"5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3","Type":"ContainerStarted","Data":"c15df03c4e70203b4b1309a98cb1c28e9f1573e8f41b2c663a2ef1227e0be3b0"} Nov 25 13:58:24 crc kubenswrapper[4702]: I1125 13:58:24.418829 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hpc6n" event={"ID":"5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3","Type":"ContainerStarted","Data":"fba75f7bbecbd5bcc0475d6ba85a37dd31b10180015aa9b767c823c23a08c8dc"} Nov 25 13:58:24 crc kubenswrapper[4702]: I1125 13:58:24.418858 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hpc6n" event={"ID":"5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3","Type":"ContainerStarted","Data":"8f2b42abbe1a096a6835163949a3d38071cd0ae7543606b50ba48082c1e42ef0"} Nov 25 13:58:24 crc kubenswrapper[4702]: I1125 13:58:24.418872 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hpc6n" event={"ID":"5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3","Type":"ContainerStarted","Data":"29f8462a8932d23ead87c19a22901f557576a4b8c5c2646b780d00f9d21043a0"} Nov 25 13:58:25 crc kubenswrapper[4702]: I1125 13:58:25.428380 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hpc6n" event={"ID":"5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3","Type":"ContainerStarted","Data":"75e9cceb72c5fb24172265fbf42fc316cc5aeab7a6838bd8879eeebaffe9e85a"} Nov 25 13:58:25 crc kubenswrapper[4702]: I1125 13:58:25.429440 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-hpc6n" Nov 25 13:58:25 crc kubenswrapper[4702]: I1125 13:58:25.455956 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-hpc6n" podStartSLOduration=6.287031998 podStartE2EDuration="14.455931038s" podCreationTimestamp="2025-11-25 13:58:11 +0000 UTC" firstStartedPulling="2025-11-25 13:58:12.335023768 +0000 UTC m=+701.497674218" lastFinishedPulling="2025-11-25 13:58:20.503922808 +0000 UTC m=+709.666573258" observedRunningTime="2025-11-25 13:58:25.452674453 +0000 UTC m=+714.615324923" watchObservedRunningTime="2025-11-25 13:58:25.455931038 +0000 UTC m=+714.618581498" Nov 25 13:58:27 crc kubenswrapper[4702]: I1125 13:58:27.193535 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-hpc6n" Nov 25 13:58:27 crc kubenswrapper[4702]: I1125 13:58:27.230964 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-hpc6n" Nov 25 13:58:32 crc kubenswrapper[4702]: I1125 13:58:32.216038 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-btnfr" Nov 25 13:58:33 crc kubenswrapper[4702]: I1125 13:58:33.573911 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-45qwr" Nov 25 13:58:40 crc kubenswrapper[4702]: I1125 13:58:40.741078 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-c5qpf"] Nov 25 13:58:40 crc kubenswrapper[4702]: I1125 13:58:40.743461 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-c5qpf" Nov 25 13:58:40 crc kubenswrapper[4702]: I1125 13:58:40.749428 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-index-dockercfg-sj74n" Nov 25 13:58:40 crc kubenswrapper[4702]: I1125 13:58:40.756961 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 25 13:58:40 crc kubenswrapper[4702]: I1125 13:58:40.757455 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 25 13:58:40 crc kubenswrapper[4702]: I1125 13:58:40.787132 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-c5qpf"] Nov 25 13:58:40 crc kubenswrapper[4702]: I1125 13:58:40.814001 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxbb7\" (UniqueName: \"kubernetes.io/projected/f3d18406-512f-42f8-acaa-ad9cc074a1a5-kube-api-access-jxbb7\") pod \"mariadb-operator-index-c5qpf\" (UID: \"f3d18406-512f-42f8-acaa-ad9cc074a1a5\") " pod="openstack-operators/mariadb-operator-index-c5qpf" Nov 25 13:58:40 crc kubenswrapper[4702]: I1125 13:58:40.915975 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxbb7\" (UniqueName: \"kubernetes.io/projected/f3d18406-512f-42f8-acaa-ad9cc074a1a5-kube-api-access-jxbb7\") pod \"mariadb-operator-index-c5qpf\" (UID: \"f3d18406-512f-42f8-acaa-ad9cc074a1a5\") " pod="openstack-operators/mariadb-operator-index-c5qpf" Nov 25 13:58:40 crc kubenswrapper[4702]: I1125 13:58:40.941875 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxbb7\" (UniqueName: \"kubernetes.io/projected/f3d18406-512f-42f8-acaa-ad9cc074a1a5-kube-api-access-jxbb7\") pod \"mariadb-operator-index-c5qpf\" (UID: \"f3d18406-512f-42f8-acaa-ad9cc074a1a5\") " pod="openstack-operators/mariadb-operator-index-c5qpf" Nov 25 13:58:41 crc kubenswrapper[4702]: I1125 13:58:41.070224 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-c5qpf" Nov 25 13:58:41 crc kubenswrapper[4702]: I1125 13:58:41.298367 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-c5qpf"] Nov 25 13:58:41 crc kubenswrapper[4702]: W1125 13:58:41.311322 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf3d18406_512f_42f8_acaa_ad9cc074a1a5.slice/crio-951f93ab951baff2a93a9fb28a357a34e3facc1dee0222afecad61cc54b0d05d WatchSource:0}: Error finding container 951f93ab951baff2a93a9fb28a357a34e3facc1dee0222afecad61cc54b0d05d: Status 404 returned error can't find the container with id 951f93ab951baff2a93a9fb28a357a34e3facc1dee0222afecad61cc54b0d05d Nov 25 13:58:41 crc kubenswrapper[4702]: I1125 13:58:41.542833 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-c5qpf" event={"ID":"f3d18406-512f-42f8-acaa-ad9cc074a1a5","Type":"ContainerStarted","Data":"951f93ab951baff2a93a9fb28a357a34e3facc1dee0222afecad61cc54b0d05d"} Nov 25 13:58:42 crc kubenswrapper[4702]: I1125 13:58:42.196406 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-hpc6n" Nov 25 13:58:43 crc kubenswrapper[4702]: I1125 13:58:43.914256 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-c5qpf"] Nov 25 13:58:44 crc kubenswrapper[4702]: I1125 13:58:44.516505 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-4p8b5"] Nov 25 13:58:44 crc kubenswrapper[4702]: I1125 13:58:44.517936 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-4p8b5" Nov 25 13:58:44 crc kubenswrapper[4702]: I1125 13:58:44.537087 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-4p8b5"] Nov 25 13:58:44 crc kubenswrapper[4702]: I1125 13:58:44.563898 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-c5qpf" event={"ID":"f3d18406-512f-42f8-acaa-ad9cc074a1a5","Type":"ContainerStarted","Data":"705b495ba0e60b087799430f8c9e12bcfcb0aa84b7c659f4ad49b63fd81e4556"} Nov 25 13:58:44 crc kubenswrapper[4702]: I1125 13:58:44.564094 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-index-c5qpf" podUID="f3d18406-512f-42f8-acaa-ad9cc074a1a5" containerName="registry-server" containerID="cri-o://705b495ba0e60b087799430f8c9e12bcfcb0aa84b7c659f4ad49b63fd81e4556" gracePeriod=2 Nov 25 13:58:44 crc kubenswrapper[4702]: I1125 13:58:44.575513 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gltj\" (UniqueName: \"kubernetes.io/projected/0eaf7ea6-c46b-4af6-af99-006a4e6e06dd-kube-api-access-5gltj\") pod \"mariadb-operator-index-4p8b5\" (UID: \"0eaf7ea6-c46b-4af6-af99-006a4e6e06dd\") " pod="openstack-operators/mariadb-operator-index-4p8b5" Nov 25 13:58:44 crc kubenswrapper[4702]: I1125 13:58:44.595180 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-c5qpf" podStartSLOduration=1.885539682 podStartE2EDuration="4.595162202s" podCreationTimestamp="2025-11-25 13:58:40 +0000 UTC" firstStartedPulling="2025-11-25 13:58:41.314706736 +0000 UTC m=+730.477357186" lastFinishedPulling="2025-11-25 13:58:44.024329266 +0000 UTC m=+733.186979706" observedRunningTime="2025-11-25 13:58:44.593654352 +0000 UTC m=+733.756304832" watchObservedRunningTime="2025-11-25 13:58:44.595162202 +0000 UTC m=+733.757812652" Nov 25 13:58:44 crc kubenswrapper[4702]: I1125 13:58:44.676985 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gltj\" (UniqueName: \"kubernetes.io/projected/0eaf7ea6-c46b-4af6-af99-006a4e6e06dd-kube-api-access-5gltj\") pod \"mariadb-operator-index-4p8b5\" (UID: \"0eaf7ea6-c46b-4af6-af99-006a4e6e06dd\") " pod="openstack-operators/mariadb-operator-index-4p8b5" Nov 25 13:58:44 crc kubenswrapper[4702]: I1125 13:58:44.705461 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gltj\" (UniqueName: \"kubernetes.io/projected/0eaf7ea6-c46b-4af6-af99-006a4e6e06dd-kube-api-access-5gltj\") pod \"mariadb-operator-index-4p8b5\" (UID: \"0eaf7ea6-c46b-4af6-af99-006a4e6e06dd\") " pod="openstack-operators/mariadb-operator-index-4p8b5" Nov 25 13:58:44 crc kubenswrapper[4702]: I1125 13:58:44.837935 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-4p8b5" Nov 25 13:58:44 crc kubenswrapper[4702]: I1125 13:58:44.915778 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-c5qpf" Nov 25 13:58:44 crc kubenswrapper[4702]: I1125 13:58:44.982869 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxbb7\" (UniqueName: \"kubernetes.io/projected/f3d18406-512f-42f8-acaa-ad9cc074a1a5-kube-api-access-jxbb7\") pod \"f3d18406-512f-42f8-acaa-ad9cc074a1a5\" (UID: \"f3d18406-512f-42f8-acaa-ad9cc074a1a5\") " Nov 25 13:58:44 crc kubenswrapper[4702]: I1125 13:58:44.987729 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3d18406-512f-42f8-acaa-ad9cc074a1a5-kube-api-access-jxbb7" (OuterVolumeSpecName: "kube-api-access-jxbb7") pod "f3d18406-512f-42f8-acaa-ad9cc074a1a5" (UID: "f3d18406-512f-42f8-acaa-ad9cc074a1a5"). InnerVolumeSpecName "kube-api-access-jxbb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:58:45 crc kubenswrapper[4702]: I1125 13:58:45.068511 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-4p8b5"] Nov 25 13:58:45 crc kubenswrapper[4702]: I1125 13:58:45.084286 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxbb7\" (UniqueName: \"kubernetes.io/projected/f3d18406-512f-42f8-acaa-ad9cc074a1a5-kube-api-access-jxbb7\") on node \"crc\" DevicePath \"\"" Nov 25 13:58:45 crc kubenswrapper[4702]: I1125 13:58:45.570711 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-4p8b5" event={"ID":"0eaf7ea6-c46b-4af6-af99-006a4e6e06dd","Type":"ContainerStarted","Data":"265f109f7801f921be8ef1e19ce768c42edba1ee1887056fb5f7a4ac61d8b0ae"} Nov 25 13:58:45 crc kubenswrapper[4702]: I1125 13:58:45.572780 4702 generic.go:334] "Generic (PLEG): container finished" podID="f3d18406-512f-42f8-acaa-ad9cc074a1a5" containerID="705b495ba0e60b087799430f8c9e12bcfcb0aa84b7c659f4ad49b63fd81e4556" exitCode=0 Nov 25 13:58:45 crc kubenswrapper[4702]: I1125 13:58:45.572841 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-c5qpf" event={"ID":"f3d18406-512f-42f8-acaa-ad9cc074a1a5","Type":"ContainerDied","Data":"705b495ba0e60b087799430f8c9e12bcfcb0aa84b7c659f4ad49b63fd81e4556"} Nov 25 13:58:45 crc kubenswrapper[4702]: I1125 13:58:45.572885 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-c5qpf" Nov 25 13:58:45 crc kubenswrapper[4702]: I1125 13:58:45.572908 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-c5qpf" event={"ID":"f3d18406-512f-42f8-acaa-ad9cc074a1a5","Type":"ContainerDied","Data":"951f93ab951baff2a93a9fb28a357a34e3facc1dee0222afecad61cc54b0d05d"} Nov 25 13:58:45 crc kubenswrapper[4702]: I1125 13:58:45.572931 4702 scope.go:117] "RemoveContainer" containerID="705b495ba0e60b087799430f8c9e12bcfcb0aa84b7c659f4ad49b63fd81e4556" Nov 25 13:58:45 crc kubenswrapper[4702]: I1125 13:58:45.604505 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-c5qpf"] Nov 25 13:58:45 crc kubenswrapper[4702]: I1125 13:58:45.604562 4702 scope.go:117] "RemoveContainer" containerID="705b495ba0e60b087799430f8c9e12bcfcb0aa84b7c659f4ad49b63fd81e4556" Nov 25 13:58:45 crc kubenswrapper[4702]: E1125 13:58:45.605160 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"705b495ba0e60b087799430f8c9e12bcfcb0aa84b7c659f4ad49b63fd81e4556\": container with ID starting with 705b495ba0e60b087799430f8c9e12bcfcb0aa84b7c659f4ad49b63fd81e4556 not found: ID does not exist" containerID="705b495ba0e60b087799430f8c9e12bcfcb0aa84b7c659f4ad49b63fd81e4556" Nov 25 13:58:45 crc kubenswrapper[4702]: I1125 13:58:45.605225 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"705b495ba0e60b087799430f8c9e12bcfcb0aa84b7c659f4ad49b63fd81e4556"} err="failed to get container status \"705b495ba0e60b087799430f8c9e12bcfcb0aa84b7c659f4ad49b63fd81e4556\": rpc error: code = NotFound desc = could not find container \"705b495ba0e60b087799430f8c9e12bcfcb0aa84b7c659f4ad49b63fd81e4556\": container with ID starting with 705b495ba0e60b087799430f8c9e12bcfcb0aa84b7c659f4ad49b63fd81e4556 not found: ID does not exist" Nov 25 13:58:45 crc kubenswrapper[4702]: I1125 13:58:45.611543 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-index-c5qpf"] Nov 25 13:58:45 crc kubenswrapper[4702]: I1125 13:58:45.811433 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3d18406-512f-42f8-acaa-ad9cc074a1a5" path="/var/lib/kubelet/pods/f3d18406-512f-42f8-acaa-ad9cc074a1a5/volumes" Nov 25 13:58:46 crc kubenswrapper[4702]: I1125 13:58:46.580018 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-4p8b5" event={"ID":"0eaf7ea6-c46b-4af6-af99-006a4e6e06dd","Type":"ContainerStarted","Data":"951881a372c4ed4d70f76f930713ec3fbd6fd2ab704045450b98f379643306a4"} Nov 25 13:58:46 crc kubenswrapper[4702]: I1125 13:58:46.597292 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-4p8b5" podStartSLOduration=1.757087205 podStartE2EDuration="2.597203114s" podCreationTimestamp="2025-11-25 13:58:44 +0000 UTC" firstStartedPulling="2025-11-25 13:58:45.08387192 +0000 UTC m=+734.246522370" lastFinishedPulling="2025-11-25 13:58:45.923987829 +0000 UTC m=+735.086638279" observedRunningTime="2025-11-25 13:58:46.595194921 +0000 UTC m=+735.757845371" watchObservedRunningTime="2025-11-25 13:58:46.597203114 +0000 UTC m=+735.759853564" Nov 25 13:58:54 crc kubenswrapper[4702]: I1125 13:58:54.839035 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/mariadb-operator-index-4p8b5" Nov 25 13:58:54 crc kubenswrapper[4702]: I1125 13:58:54.839994 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-index-4p8b5" Nov 25 13:58:54 crc kubenswrapper[4702]: I1125 13:58:54.870453 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/mariadb-operator-index-4p8b5" Nov 25 13:58:55 crc kubenswrapper[4702]: I1125 13:58:55.665977 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-index-4p8b5" Nov 25 13:59:01 crc kubenswrapper[4702]: I1125 13:59:01.913186 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9"] Nov 25 13:59:01 crc kubenswrapper[4702]: E1125 13:59:01.914057 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3d18406-512f-42f8-acaa-ad9cc074a1a5" containerName="registry-server" Nov 25 13:59:01 crc kubenswrapper[4702]: I1125 13:59:01.914073 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3d18406-512f-42f8-acaa-ad9cc074a1a5" containerName="registry-server" Nov 25 13:59:01 crc kubenswrapper[4702]: I1125 13:59:01.914189 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3d18406-512f-42f8-acaa-ad9cc074a1a5" containerName="registry-server" Nov 25 13:59:01 crc kubenswrapper[4702]: I1125 13:59:01.914988 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9" Nov 25 13:59:01 crc kubenswrapper[4702]: I1125 13:59:01.917916 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-6r4jq" Nov 25 13:59:01 crc kubenswrapper[4702]: I1125 13:59:01.939224 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9"] Nov 25 13:59:02 crc kubenswrapper[4702]: I1125 13:59:02.027894 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8765d4d8-ce9f-4801-8d84-1badefac8c10-bundle\") pod \"c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9\" (UID: \"8765d4d8-ce9f-4801-8d84-1badefac8c10\") " pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9" Nov 25 13:59:02 crc kubenswrapper[4702]: I1125 13:59:02.027973 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8765d4d8-ce9f-4801-8d84-1badefac8c10-util\") pod \"c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9\" (UID: \"8765d4d8-ce9f-4801-8d84-1badefac8c10\") " pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9" Nov 25 13:59:02 crc kubenswrapper[4702]: I1125 13:59:02.028160 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrtgr\" (UniqueName: \"kubernetes.io/projected/8765d4d8-ce9f-4801-8d84-1badefac8c10-kube-api-access-vrtgr\") pod \"c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9\" (UID: \"8765d4d8-ce9f-4801-8d84-1badefac8c10\") " pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9" Nov 25 13:59:02 crc kubenswrapper[4702]: I1125 13:59:02.129741 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8765d4d8-ce9f-4801-8d84-1badefac8c10-bundle\") pod \"c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9\" (UID: \"8765d4d8-ce9f-4801-8d84-1badefac8c10\") " pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9" Nov 25 13:59:02 crc kubenswrapper[4702]: I1125 13:59:02.129818 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8765d4d8-ce9f-4801-8d84-1badefac8c10-util\") pod \"c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9\" (UID: \"8765d4d8-ce9f-4801-8d84-1badefac8c10\") " pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9" Nov 25 13:59:02 crc kubenswrapper[4702]: I1125 13:59:02.129939 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrtgr\" (UniqueName: \"kubernetes.io/projected/8765d4d8-ce9f-4801-8d84-1badefac8c10-kube-api-access-vrtgr\") pod \"c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9\" (UID: \"8765d4d8-ce9f-4801-8d84-1badefac8c10\") " pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9" Nov 25 13:59:02 crc kubenswrapper[4702]: I1125 13:59:02.130770 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8765d4d8-ce9f-4801-8d84-1badefac8c10-util\") pod \"c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9\" (UID: \"8765d4d8-ce9f-4801-8d84-1badefac8c10\") " pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9" Nov 25 13:59:02 crc kubenswrapper[4702]: I1125 13:59:02.130789 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8765d4d8-ce9f-4801-8d84-1badefac8c10-bundle\") pod \"c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9\" (UID: \"8765d4d8-ce9f-4801-8d84-1badefac8c10\") " pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9" Nov 25 13:59:02 crc kubenswrapper[4702]: I1125 13:59:02.152886 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrtgr\" (UniqueName: \"kubernetes.io/projected/8765d4d8-ce9f-4801-8d84-1badefac8c10-kube-api-access-vrtgr\") pod \"c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9\" (UID: \"8765d4d8-ce9f-4801-8d84-1badefac8c10\") " pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9" Nov 25 13:59:02 crc kubenswrapper[4702]: I1125 13:59:02.242277 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9" Nov 25 13:59:02 crc kubenswrapper[4702]: I1125 13:59:02.474060 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9"] Nov 25 13:59:02 crc kubenswrapper[4702]: I1125 13:59:02.678424 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9" event={"ID":"8765d4d8-ce9f-4801-8d84-1badefac8c10","Type":"ContainerStarted","Data":"0e78daeca36234f724db26a7e7e9ff4b54e1075335f262aff17c4bca786c3201"} Nov 25 13:59:03 crc kubenswrapper[4702]: I1125 13:59:03.689072 4702 generic.go:334] "Generic (PLEG): container finished" podID="8765d4d8-ce9f-4801-8d84-1badefac8c10" containerID="a08092c7f059bdd9cee27250e7276c8d0be39711ca80999433a7aabc8fb4fd83" exitCode=0 Nov 25 13:59:03 crc kubenswrapper[4702]: I1125 13:59:03.689427 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9" event={"ID":"8765d4d8-ce9f-4801-8d84-1badefac8c10","Type":"ContainerDied","Data":"a08092c7f059bdd9cee27250e7276c8d0be39711ca80999433a7aabc8fb4fd83"} Nov 25 13:59:05 crc kubenswrapper[4702]: I1125 13:59:05.702456 4702 generic.go:334] "Generic (PLEG): container finished" podID="8765d4d8-ce9f-4801-8d84-1badefac8c10" containerID="edacee0121d986e172e9e8c098858373b95722ff9408eb4d6ac16d2e0af6997a" exitCode=0 Nov 25 13:59:05 crc kubenswrapper[4702]: I1125 13:59:05.702564 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9" event={"ID":"8765d4d8-ce9f-4801-8d84-1badefac8c10","Type":"ContainerDied","Data":"edacee0121d986e172e9e8c098858373b95722ff9408eb4d6ac16d2e0af6997a"} Nov 25 13:59:06 crc kubenswrapper[4702]: I1125 13:59:06.711561 4702 generic.go:334] "Generic (PLEG): container finished" podID="8765d4d8-ce9f-4801-8d84-1badefac8c10" containerID="48ecc429b75d75931c252d456cd377756168e1049def787ffb29bd20f32da07f" exitCode=0 Nov 25 13:59:06 crc kubenswrapper[4702]: I1125 13:59:06.711617 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9" event={"ID":"8765d4d8-ce9f-4801-8d84-1badefac8c10","Type":"ContainerDied","Data":"48ecc429b75d75931c252d456cd377756168e1049def787ffb29bd20f32da07f"} Nov 25 13:59:07 crc kubenswrapper[4702]: I1125 13:59:07.958201 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9" Nov 25 13:59:08 crc kubenswrapper[4702]: I1125 13:59:08.003791 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrtgr\" (UniqueName: \"kubernetes.io/projected/8765d4d8-ce9f-4801-8d84-1badefac8c10-kube-api-access-vrtgr\") pod \"8765d4d8-ce9f-4801-8d84-1badefac8c10\" (UID: \"8765d4d8-ce9f-4801-8d84-1badefac8c10\") " Nov 25 13:59:08 crc kubenswrapper[4702]: I1125 13:59:08.003937 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8765d4d8-ce9f-4801-8d84-1badefac8c10-util\") pod \"8765d4d8-ce9f-4801-8d84-1badefac8c10\" (UID: \"8765d4d8-ce9f-4801-8d84-1badefac8c10\") " Nov 25 13:59:08 crc kubenswrapper[4702]: I1125 13:59:08.003972 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8765d4d8-ce9f-4801-8d84-1badefac8c10-bundle\") pod \"8765d4d8-ce9f-4801-8d84-1badefac8c10\" (UID: \"8765d4d8-ce9f-4801-8d84-1badefac8c10\") " Nov 25 13:59:08 crc kubenswrapper[4702]: I1125 13:59:08.005541 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8765d4d8-ce9f-4801-8d84-1badefac8c10-bundle" (OuterVolumeSpecName: "bundle") pod "8765d4d8-ce9f-4801-8d84-1badefac8c10" (UID: "8765d4d8-ce9f-4801-8d84-1badefac8c10"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:59:08 crc kubenswrapper[4702]: I1125 13:59:08.012451 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8765d4d8-ce9f-4801-8d84-1badefac8c10-kube-api-access-vrtgr" (OuterVolumeSpecName: "kube-api-access-vrtgr") pod "8765d4d8-ce9f-4801-8d84-1badefac8c10" (UID: "8765d4d8-ce9f-4801-8d84-1badefac8c10"). InnerVolumeSpecName "kube-api-access-vrtgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:59:08 crc kubenswrapper[4702]: I1125 13:59:08.105137 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrtgr\" (UniqueName: \"kubernetes.io/projected/8765d4d8-ce9f-4801-8d84-1badefac8c10-kube-api-access-vrtgr\") on node \"crc\" DevicePath \"\"" Nov 25 13:59:08 crc kubenswrapper[4702]: I1125 13:59:08.105622 4702 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8765d4d8-ce9f-4801-8d84-1badefac8c10-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 13:59:08 crc kubenswrapper[4702]: I1125 13:59:08.338157 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8765d4d8-ce9f-4801-8d84-1badefac8c10-util" (OuterVolumeSpecName: "util") pod "8765d4d8-ce9f-4801-8d84-1badefac8c10" (UID: "8765d4d8-ce9f-4801-8d84-1badefac8c10"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:59:08 crc kubenswrapper[4702]: I1125 13:59:08.408725 4702 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8765d4d8-ce9f-4801-8d84-1badefac8c10-util\") on node \"crc\" DevicePath \"\"" Nov 25 13:59:08 crc kubenswrapper[4702]: I1125 13:59:08.729312 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9" event={"ID":"8765d4d8-ce9f-4801-8d84-1badefac8c10","Type":"ContainerDied","Data":"0e78daeca36234f724db26a7e7e9ff4b54e1075335f262aff17c4bca786c3201"} Nov 25 13:59:08 crc kubenswrapper[4702]: I1125 13:59:08.729360 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e78daeca36234f724db26a7e7e9ff4b54e1075335f262aff17c4bca786c3201" Nov 25 13:59:08 crc kubenswrapper[4702]: I1125 13:59:08.729435 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9" Nov 25 13:59:14 crc kubenswrapper[4702]: I1125 13:59:14.850671 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-64cc8659b5-67ssg"] Nov 25 13:59:14 crc kubenswrapper[4702]: E1125 13:59:14.851184 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8765d4d8-ce9f-4801-8d84-1badefac8c10" containerName="extract" Nov 25 13:59:14 crc kubenswrapper[4702]: I1125 13:59:14.851198 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="8765d4d8-ce9f-4801-8d84-1badefac8c10" containerName="extract" Nov 25 13:59:14 crc kubenswrapper[4702]: E1125 13:59:14.851208 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8765d4d8-ce9f-4801-8d84-1badefac8c10" containerName="pull" Nov 25 13:59:14 crc kubenswrapper[4702]: I1125 13:59:14.851214 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="8765d4d8-ce9f-4801-8d84-1badefac8c10" containerName="pull" Nov 25 13:59:14 crc kubenswrapper[4702]: E1125 13:59:14.851262 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8765d4d8-ce9f-4801-8d84-1badefac8c10" containerName="util" Nov 25 13:59:14 crc kubenswrapper[4702]: I1125 13:59:14.851270 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="8765d4d8-ce9f-4801-8d84-1badefac8c10" containerName="util" Nov 25 13:59:14 crc kubenswrapper[4702]: I1125 13:59:14.851370 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="8765d4d8-ce9f-4801-8d84-1badefac8c10" containerName="extract" Nov 25 13:59:14 crc kubenswrapper[4702]: I1125 13:59:14.851739 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-64cc8659b5-67ssg" Nov 25 13:59:14 crc kubenswrapper[4702]: I1125 13:59:14.854057 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-kqzb7" Nov 25 13:59:14 crc kubenswrapper[4702]: I1125 13:59:14.854307 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-service-cert" Nov 25 13:59:14 crc kubenswrapper[4702]: I1125 13:59:14.854456 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 25 13:59:14 crc kubenswrapper[4702]: I1125 13:59:14.870892 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-64cc8659b5-67ssg"] Nov 25 13:59:14 crc kubenswrapper[4702]: I1125 13:59:14.917706 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9817b60d-cfac-44d3-b102-ec3a5670fbc2-webhook-cert\") pod \"mariadb-operator-controller-manager-64cc8659b5-67ssg\" (UID: \"9817b60d-cfac-44d3-b102-ec3a5670fbc2\") " pod="openstack-operators/mariadb-operator-controller-manager-64cc8659b5-67ssg" Nov 25 13:59:14 crc kubenswrapper[4702]: I1125 13:59:14.917776 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq28j\" (UniqueName: \"kubernetes.io/projected/9817b60d-cfac-44d3-b102-ec3a5670fbc2-kube-api-access-lq28j\") pod \"mariadb-operator-controller-manager-64cc8659b5-67ssg\" (UID: \"9817b60d-cfac-44d3-b102-ec3a5670fbc2\") " pod="openstack-operators/mariadb-operator-controller-manager-64cc8659b5-67ssg" Nov 25 13:59:14 crc kubenswrapper[4702]: I1125 13:59:14.917799 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9817b60d-cfac-44d3-b102-ec3a5670fbc2-apiservice-cert\") pod \"mariadb-operator-controller-manager-64cc8659b5-67ssg\" (UID: \"9817b60d-cfac-44d3-b102-ec3a5670fbc2\") " pod="openstack-operators/mariadb-operator-controller-manager-64cc8659b5-67ssg" Nov 25 13:59:15 crc kubenswrapper[4702]: I1125 13:59:15.019117 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9817b60d-cfac-44d3-b102-ec3a5670fbc2-webhook-cert\") pod \"mariadb-operator-controller-manager-64cc8659b5-67ssg\" (UID: \"9817b60d-cfac-44d3-b102-ec3a5670fbc2\") " pod="openstack-operators/mariadb-operator-controller-manager-64cc8659b5-67ssg" Nov 25 13:59:15 crc kubenswrapper[4702]: I1125 13:59:15.019454 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq28j\" (UniqueName: \"kubernetes.io/projected/9817b60d-cfac-44d3-b102-ec3a5670fbc2-kube-api-access-lq28j\") pod \"mariadb-operator-controller-manager-64cc8659b5-67ssg\" (UID: \"9817b60d-cfac-44d3-b102-ec3a5670fbc2\") " pod="openstack-operators/mariadb-operator-controller-manager-64cc8659b5-67ssg" Nov 25 13:59:15 crc kubenswrapper[4702]: I1125 13:59:15.019541 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9817b60d-cfac-44d3-b102-ec3a5670fbc2-apiservice-cert\") pod \"mariadb-operator-controller-manager-64cc8659b5-67ssg\" (UID: \"9817b60d-cfac-44d3-b102-ec3a5670fbc2\") " pod="openstack-operators/mariadb-operator-controller-manager-64cc8659b5-67ssg" Nov 25 13:59:15 crc kubenswrapper[4702]: I1125 13:59:15.025398 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9817b60d-cfac-44d3-b102-ec3a5670fbc2-webhook-cert\") pod \"mariadb-operator-controller-manager-64cc8659b5-67ssg\" (UID: \"9817b60d-cfac-44d3-b102-ec3a5670fbc2\") " pod="openstack-operators/mariadb-operator-controller-manager-64cc8659b5-67ssg" Nov 25 13:59:15 crc kubenswrapper[4702]: I1125 13:59:15.026989 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9817b60d-cfac-44d3-b102-ec3a5670fbc2-apiservice-cert\") pod \"mariadb-operator-controller-manager-64cc8659b5-67ssg\" (UID: \"9817b60d-cfac-44d3-b102-ec3a5670fbc2\") " pod="openstack-operators/mariadb-operator-controller-manager-64cc8659b5-67ssg" Nov 25 13:59:15 crc kubenswrapper[4702]: I1125 13:59:15.036924 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq28j\" (UniqueName: \"kubernetes.io/projected/9817b60d-cfac-44d3-b102-ec3a5670fbc2-kube-api-access-lq28j\") pod \"mariadb-operator-controller-manager-64cc8659b5-67ssg\" (UID: \"9817b60d-cfac-44d3-b102-ec3a5670fbc2\") " pod="openstack-operators/mariadb-operator-controller-manager-64cc8659b5-67ssg" Nov 25 13:59:15 crc kubenswrapper[4702]: I1125 13:59:15.176424 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-64cc8659b5-67ssg" Nov 25 13:59:15 crc kubenswrapper[4702]: I1125 13:59:15.477855 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-64cc8659b5-67ssg"] Nov 25 13:59:15 crc kubenswrapper[4702]: I1125 13:59:15.773467 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-64cc8659b5-67ssg" event={"ID":"9817b60d-cfac-44d3-b102-ec3a5670fbc2","Type":"ContainerStarted","Data":"2d009812b45cce74b6dc916d0ea6a02a3a99eb3bcd06bacb1ac7cfaf874cb70a"} Nov 25 13:59:20 crc kubenswrapper[4702]: I1125 13:59:20.807624 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-64cc8659b5-67ssg" event={"ID":"9817b60d-cfac-44d3-b102-ec3a5670fbc2","Type":"ContainerStarted","Data":"07d481ba5cdec1c5ba77cba813a258b42fdb21b815455aa90d916dbfc185dd2c"} Nov 25 13:59:20 crc kubenswrapper[4702]: I1125 13:59:20.808283 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-64cc8659b5-67ssg" Nov 25 13:59:25 crc kubenswrapper[4702]: I1125 13:59:25.182716 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-64cc8659b5-67ssg" Nov 25 13:59:25 crc kubenswrapper[4702]: I1125 13:59:25.205034 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-64cc8659b5-67ssg" podStartSLOduration=6.609314748 podStartE2EDuration="11.205013326s" podCreationTimestamp="2025-11-25 13:59:14 +0000 UTC" firstStartedPulling="2025-11-25 13:59:15.486029533 +0000 UTC m=+764.648679983" lastFinishedPulling="2025-11-25 13:59:20.081728111 +0000 UTC m=+769.244378561" observedRunningTime="2025-11-25 13:59:20.826735741 +0000 UTC m=+769.989386191" watchObservedRunningTime="2025-11-25 13:59:25.205013326 +0000 UTC m=+774.367663796" Nov 25 13:59:30 crc kubenswrapper[4702]: I1125 13:59:30.389870 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-mgnpn"] Nov 25 13:59:30 crc kubenswrapper[4702]: I1125 13:59:30.390931 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-mgnpn" Nov 25 13:59:30 crc kubenswrapper[4702]: I1125 13:59:30.392778 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-index-dockercfg-bjwjp" Nov 25 13:59:30 crc kubenswrapper[4702]: I1125 13:59:30.399755 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-mgnpn"] Nov 25 13:59:30 crc kubenswrapper[4702]: I1125 13:59:30.433931 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfs9v\" (UniqueName: \"kubernetes.io/projected/bf684cff-ed36-404a-8eba-acdb4cac164d-kube-api-access-sfs9v\") pod \"infra-operator-index-mgnpn\" (UID: \"bf684cff-ed36-404a-8eba-acdb4cac164d\") " pod="openstack-operators/infra-operator-index-mgnpn" Nov 25 13:59:30 crc kubenswrapper[4702]: I1125 13:59:30.535496 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfs9v\" (UniqueName: \"kubernetes.io/projected/bf684cff-ed36-404a-8eba-acdb4cac164d-kube-api-access-sfs9v\") pod \"infra-operator-index-mgnpn\" (UID: \"bf684cff-ed36-404a-8eba-acdb4cac164d\") " pod="openstack-operators/infra-operator-index-mgnpn" Nov 25 13:59:30 crc kubenswrapper[4702]: I1125 13:59:30.555089 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfs9v\" (UniqueName: \"kubernetes.io/projected/bf684cff-ed36-404a-8eba-acdb4cac164d-kube-api-access-sfs9v\") pod \"infra-operator-index-mgnpn\" (UID: \"bf684cff-ed36-404a-8eba-acdb4cac164d\") " pod="openstack-operators/infra-operator-index-mgnpn" Nov 25 13:59:30 crc kubenswrapper[4702]: I1125 13:59:30.710311 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-mgnpn" Nov 25 13:59:33 crc kubenswrapper[4702]: I1125 13:59:33.012673 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-mgnpn"] Nov 25 13:59:33 crc kubenswrapper[4702]: W1125 13:59:33.028746 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf684cff_ed36_404a_8eba_acdb4cac164d.slice/crio-d44bbdbe9f2cbb1a4ec378aa5023f365cf9894fd5b37d373c4b39efbcb6c8431 WatchSource:0}: Error finding container d44bbdbe9f2cbb1a4ec378aa5023f365cf9894fd5b37d373c4b39efbcb6c8431: Status 404 returned error can't find the container with id d44bbdbe9f2cbb1a4ec378aa5023f365cf9894fd5b37d373c4b39efbcb6c8431 Nov 25 13:59:33 crc kubenswrapper[4702]: I1125 13:59:33.891203 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-mgnpn" event={"ID":"bf684cff-ed36-404a-8eba-acdb4cac164d","Type":"ContainerStarted","Data":"d44bbdbe9f2cbb1a4ec378aa5023f365cf9894fd5b37d373c4b39efbcb6c8431"} Nov 25 13:59:33 crc kubenswrapper[4702]: I1125 13:59:33.963077 4702 patch_prober.go:28] interesting pod/machine-config-daemon-qqr76 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 13:59:33 crc kubenswrapper[4702]: I1125 13:59:33.963180 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 13:59:34 crc kubenswrapper[4702]: I1125 13:59:34.901754 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-mgnpn" event={"ID":"bf684cff-ed36-404a-8eba-acdb4cac164d","Type":"ContainerStarted","Data":"f66416e1b54881b2aaa611200ea46cfc6eac9534ab6cc4b9f1e9241ac940f5d9"} Nov 25 13:59:34 crc kubenswrapper[4702]: I1125 13:59:34.928645 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-mgnpn" podStartSLOduration=3.467811538 podStartE2EDuration="4.92861832s" podCreationTimestamp="2025-11-25 13:59:30 +0000 UTC" firstStartedPulling="2025-11-25 13:59:33.030291416 +0000 UTC m=+782.192941866" lastFinishedPulling="2025-11-25 13:59:34.491098198 +0000 UTC m=+783.653748648" observedRunningTime="2025-11-25 13:59:34.919082888 +0000 UTC m=+784.081733338" watchObservedRunningTime="2025-11-25 13:59:34.92861832 +0000 UTC m=+784.091268770" Nov 25 13:59:40 crc kubenswrapper[4702]: I1125 13:59:40.711161 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/infra-operator-index-mgnpn" Nov 25 13:59:40 crc kubenswrapper[4702]: I1125 13:59:40.712085 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-index-mgnpn" Nov 25 13:59:40 crc kubenswrapper[4702]: I1125 13:59:40.745951 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/infra-operator-index-mgnpn" Nov 25 13:59:40 crc kubenswrapper[4702]: I1125 13:59:40.964934 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-index-mgnpn" Nov 25 13:59:42 crc kubenswrapper[4702]: I1125 13:59:42.435050 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g"] Nov 25 13:59:42 crc kubenswrapper[4702]: I1125 13:59:42.436615 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g" Nov 25 13:59:42 crc kubenswrapper[4702]: I1125 13:59:42.445536 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-6r4jq" Nov 25 13:59:42 crc kubenswrapper[4702]: I1125 13:59:42.446007 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g"] Nov 25 13:59:42 crc kubenswrapper[4702]: I1125 13:59:42.501676 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqcwf\" (UniqueName: \"kubernetes.io/projected/cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba-kube-api-access-vqcwf\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g\" (UID: \"cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g" Nov 25 13:59:42 crc kubenswrapper[4702]: I1125 13:59:42.501722 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba-util\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g\" (UID: \"cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g" Nov 25 13:59:42 crc kubenswrapper[4702]: I1125 13:59:42.501781 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba-bundle\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g\" (UID: \"cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g" Nov 25 13:59:42 crc kubenswrapper[4702]: I1125 13:59:42.603699 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqcwf\" (UniqueName: \"kubernetes.io/projected/cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba-kube-api-access-vqcwf\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g\" (UID: \"cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g" Nov 25 13:59:42 crc kubenswrapper[4702]: I1125 13:59:42.603755 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba-util\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g\" (UID: \"cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g" Nov 25 13:59:42 crc kubenswrapper[4702]: I1125 13:59:42.603803 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba-bundle\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g\" (UID: \"cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g" Nov 25 13:59:42 crc kubenswrapper[4702]: I1125 13:59:42.604383 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba-bundle\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g\" (UID: \"cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g" Nov 25 13:59:42 crc kubenswrapper[4702]: I1125 13:59:42.604596 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba-util\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g\" (UID: \"cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g" Nov 25 13:59:42 crc kubenswrapper[4702]: I1125 13:59:42.626038 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqcwf\" (UniqueName: \"kubernetes.io/projected/cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba-kube-api-access-vqcwf\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g\" (UID: \"cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g" Nov 25 13:59:42 crc kubenswrapper[4702]: I1125 13:59:42.761297 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g" Nov 25 13:59:43 crc kubenswrapper[4702]: I1125 13:59:43.233495 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g"] Nov 25 13:59:43 crc kubenswrapper[4702]: I1125 13:59:43.957796 4702 generic.go:334] "Generic (PLEG): container finished" podID="cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba" containerID="09a6d0c4bf9938b64f246c7b061706f07e4635987769e8b4ed997b3c3341881a" exitCode=0 Nov 25 13:59:43 crc kubenswrapper[4702]: I1125 13:59:43.957935 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g" event={"ID":"cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba","Type":"ContainerDied","Data":"09a6d0c4bf9938b64f246c7b061706f07e4635987769e8b4ed997b3c3341881a"} Nov 25 13:59:43 crc kubenswrapper[4702]: I1125 13:59:43.958159 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g" event={"ID":"cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba","Type":"ContainerStarted","Data":"cc7b0b3be238a60e4ce1432a20e9af44e7b0339850d7942fdf15b621d0f8354e"} Nov 25 13:59:46 crc kubenswrapper[4702]: I1125 13:59:46.979372 4702 generic.go:334] "Generic (PLEG): container finished" podID="cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba" containerID="fa11d96fafa80170fc93dc94d5a3a914b0bbbd2fd51e0e5607e68abb483906ff" exitCode=0 Nov 25 13:59:46 crc kubenswrapper[4702]: I1125 13:59:46.979540 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g" event={"ID":"cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba","Type":"ContainerDied","Data":"fa11d96fafa80170fc93dc94d5a3a914b0bbbd2fd51e0e5607e68abb483906ff"} Nov 25 13:59:47 crc kubenswrapper[4702]: I1125 13:59:47.991247 4702 generic.go:334] "Generic (PLEG): container finished" podID="cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba" containerID="75b876a101dc7ea85f6034dac9234adffaeef88a4ce98325581f4a5f372d518b" exitCode=0 Nov 25 13:59:47 crc kubenswrapper[4702]: I1125 13:59:47.991362 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g" event={"ID":"cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba","Type":"ContainerDied","Data":"75b876a101dc7ea85f6034dac9234adffaeef88a4ce98325581f4a5f372d518b"} Nov 25 13:59:49 crc kubenswrapper[4702]: I1125 13:59:49.304998 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g" Nov 25 13:59:49 crc kubenswrapper[4702]: I1125 13:59:49.402836 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba-util\") pod \"cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba\" (UID: \"cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba\") " Nov 25 13:59:49 crc kubenswrapper[4702]: I1125 13:59:49.402981 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba-bundle\") pod \"cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba\" (UID: \"cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba\") " Nov 25 13:59:49 crc kubenswrapper[4702]: I1125 13:59:49.403258 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqcwf\" (UniqueName: \"kubernetes.io/projected/cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba-kube-api-access-vqcwf\") pod \"cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba\" (UID: \"cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba\") " Nov 25 13:59:49 crc kubenswrapper[4702]: I1125 13:59:49.404531 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba-bundle" (OuterVolumeSpecName: "bundle") pod "cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba" (UID: "cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:59:49 crc kubenswrapper[4702]: I1125 13:59:49.410623 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba-kube-api-access-vqcwf" (OuterVolumeSpecName: "kube-api-access-vqcwf") pod "cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba" (UID: "cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba"). InnerVolumeSpecName "kube-api-access-vqcwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 13:59:49 crc kubenswrapper[4702]: I1125 13:59:49.506054 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqcwf\" (UniqueName: \"kubernetes.io/projected/cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba-kube-api-access-vqcwf\") on node \"crc\" DevicePath \"\"" Nov 25 13:59:49 crc kubenswrapper[4702]: I1125 13:59:49.506132 4702 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 13:59:49 crc kubenswrapper[4702]: I1125 13:59:49.575906 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba-util" (OuterVolumeSpecName: "util") pod "cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba" (UID: "cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 13:59:49 crc kubenswrapper[4702]: I1125 13:59:49.607879 4702 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba-util\") on node \"crc\" DevicePath \"\"" Nov 25 13:59:50 crc kubenswrapper[4702]: I1125 13:59:50.004928 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g" event={"ID":"cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba","Type":"ContainerDied","Data":"cc7b0b3be238a60e4ce1432a20e9af44e7b0339850d7942fdf15b621d0f8354e"} Nov 25 13:59:50 crc kubenswrapper[4702]: I1125 13:59:50.004963 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc7b0b3be238a60e4ce1432a20e9af44e7b0339850d7942fdf15b621d0f8354e" Nov 25 13:59:50 crc kubenswrapper[4702]: I1125 13:59:50.004989 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g" Nov 25 13:59:55 crc kubenswrapper[4702]: I1125 13:59:55.041115 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-79898f8746-rfpp7"] Nov 25 13:59:55 crc kubenswrapper[4702]: E1125 13:59:55.041989 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba" containerName="pull" Nov 25 13:59:55 crc kubenswrapper[4702]: I1125 13:59:55.042009 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba" containerName="pull" Nov 25 13:59:55 crc kubenswrapper[4702]: E1125 13:59:55.042031 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba" containerName="util" Nov 25 13:59:55 crc kubenswrapper[4702]: I1125 13:59:55.042040 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba" containerName="util" Nov 25 13:59:55 crc kubenswrapper[4702]: E1125 13:59:55.042054 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba" containerName="extract" Nov 25 13:59:55 crc kubenswrapper[4702]: I1125 13:59:55.042061 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba" containerName="extract" Nov 25 13:59:55 crc kubenswrapper[4702]: I1125 13:59:55.042177 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba" containerName="extract" Nov 25 13:59:55 crc kubenswrapper[4702]: I1125 13:59:55.043708 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79898f8746-rfpp7" Nov 25 13:59:55 crc kubenswrapper[4702]: I1125 13:59:55.059901 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-service-cert" Nov 25 13:59:55 crc kubenswrapper[4702]: I1125 13:59:55.060014 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-4qbks" Nov 25 13:59:55 crc kubenswrapper[4702]: I1125 13:59:55.093785 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79898f8746-rfpp7"] Nov 25 13:59:55 crc kubenswrapper[4702]: I1125 13:59:55.097994 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e13b9e67-df17-4851-a3c5-5037b994637d-webhook-cert\") pod \"infra-operator-controller-manager-79898f8746-rfpp7\" (UID: \"e13b9e67-df17-4851-a3c5-5037b994637d\") " pod="openstack-operators/infra-operator-controller-manager-79898f8746-rfpp7" Nov 25 13:59:55 crc kubenswrapper[4702]: I1125 13:59:55.098079 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzjbj\" (UniqueName: \"kubernetes.io/projected/e13b9e67-df17-4851-a3c5-5037b994637d-kube-api-access-jzjbj\") pod \"infra-operator-controller-manager-79898f8746-rfpp7\" (UID: \"e13b9e67-df17-4851-a3c5-5037b994637d\") " pod="openstack-operators/infra-operator-controller-manager-79898f8746-rfpp7" Nov 25 13:59:55 crc kubenswrapper[4702]: I1125 13:59:55.098127 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e13b9e67-df17-4851-a3c5-5037b994637d-apiservice-cert\") pod \"infra-operator-controller-manager-79898f8746-rfpp7\" (UID: \"e13b9e67-df17-4851-a3c5-5037b994637d\") " pod="openstack-operators/infra-operator-controller-manager-79898f8746-rfpp7" Nov 25 13:59:55 crc kubenswrapper[4702]: I1125 13:59:55.199668 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e13b9e67-df17-4851-a3c5-5037b994637d-apiservice-cert\") pod \"infra-operator-controller-manager-79898f8746-rfpp7\" (UID: \"e13b9e67-df17-4851-a3c5-5037b994637d\") " pod="openstack-operators/infra-operator-controller-manager-79898f8746-rfpp7" Nov 25 13:59:55 crc kubenswrapper[4702]: I1125 13:59:55.199745 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e13b9e67-df17-4851-a3c5-5037b994637d-webhook-cert\") pod \"infra-operator-controller-manager-79898f8746-rfpp7\" (UID: \"e13b9e67-df17-4851-a3c5-5037b994637d\") " pod="openstack-operators/infra-operator-controller-manager-79898f8746-rfpp7" Nov 25 13:59:55 crc kubenswrapper[4702]: I1125 13:59:55.199782 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzjbj\" (UniqueName: \"kubernetes.io/projected/e13b9e67-df17-4851-a3c5-5037b994637d-kube-api-access-jzjbj\") pod \"infra-operator-controller-manager-79898f8746-rfpp7\" (UID: \"e13b9e67-df17-4851-a3c5-5037b994637d\") " pod="openstack-operators/infra-operator-controller-manager-79898f8746-rfpp7" Nov 25 13:59:55 crc kubenswrapper[4702]: I1125 13:59:55.205676 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e13b9e67-df17-4851-a3c5-5037b994637d-apiservice-cert\") pod \"infra-operator-controller-manager-79898f8746-rfpp7\" (UID: \"e13b9e67-df17-4851-a3c5-5037b994637d\") " pod="openstack-operators/infra-operator-controller-manager-79898f8746-rfpp7" Nov 25 13:59:55 crc kubenswrapper[4702]: I1125 13:59:55.219677 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e13b9e67-df17-4851-a3c5-5037b994637d-webhook-cert\") pod \"infra-operator-controller-manager-79898f8746-rfpp7\" (UID: \"e13b9e67-df17-4851-a3c5-5037b994637d\") " pod="openstack-operators/infra-operator-controller-manager-79898f8746-rfpp7" Nov 25 13:59:55 crc kubenswrapper[4702]: I1125 13:59:55.222851 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzjbj\" (UniqueName: \"kubernetes.io/projected/e13b9e67-df17-4851-a3c5-5037b994637d-kube-api-access-jzjbj\") pod \"infra-operator-controller-manager-79898f8746-rfpp7\" (UID: \"e13b9e67-df17-4851-a3c5-5037b994637d\") " pod="openstack-operators/infra-operator-controller-manager-79898f8746-rfpp7" Nov 25 13:59:55 crc kubenswrapper[4702]: I1125 13:59:55.381044 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79898f8746-rfpp7" Nov 25 13:59:55 crc kubenswrapper[4702]: I1125 13:59:55.797018 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79898f8746-rfpp7"] Nov 25 13:59:55 crc kubenswrapper[4702]: W1125 13:59:55.804325 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode13b9e67_df17_4851_a3c5_5037b994637d.slice/crio-bc45407b9769afd3805d7819ba109ba5b5bf99b57436553a28a4def0829f3fa7 WatchSource:0}: Error finding container bc45407b9769afd3805d7819ba109ba5b5bf99b57436553a28a4def0829f3fa7: Status 404 returned error can't find the container with id bc45407b9769afd3805d7819ba109ba5b5bf99b57436553a28a4def0829f3fa7 Nov 25 13:59:56 crc kubenswrapper[4702]: I1125 13:59:56.046612 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79898f8746-rfpp7" event={"ID":"e13b9e67-df17-4851-a3c5-5037b994637d","Type":"ContainerStarted","Data":"bc45407b9769afd3805d7819ba109ba5b5bf99b57436553a28a4def0829f3fa7"} Nov 25 14:00:00 crc kubenswrapper[4702]: I1125 14:00:00.076675 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79898f8746-rfpp7" event={"ID":"e13b9e67-df17-4851-a3c5-5037b994637d","Type":"ContainerStarted","Data":"1cdc51de5be45f83a96dbb0e74a3967d74d9d32d0da704751791c94b55d6728a"} Nov 25 14:00:00 crc kubenswrapper[4702]: I1125 14:00:00.178742 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401320-jbp7s"] Nov 25 14:00:00 crc kubenswrapper[4702]: I1125 14:00:00.179787 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401320-jbp7s" Nov 25 14:00:00 crc kubenswrapper[4702]: I1125 14:00:00.184775 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 14:00:00 crc kubenswrapper[4702]: I1125 14:00:00.187087 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 14:00:00 crc kubenswrapper[4702]: I1125 14:00:00.191096 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401320-jbp7s"] Nov 25 14:00:00 crc kubenswrapper[4702]: I1125 14:00:00.272402 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpbw8\" (UniqueName: \"kubernetes.io/projected/5f2bd1d5-4ceb-4191-9f8f-255f157acecf-kube-api-access-bpbw8\") pod \"collect-profiles-29401320-jbp7s\" (UID: \"5f2bd1d5-4ceb-4191-9f8f-255f157acecf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401320-jbp7s" Nov 25 14:00:00 crc kubenswrapper[4702]: I1125 14:00:00.272634 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5f2bd1d5-4ceb-4191-9f8f-255f157acecf-config-volume\") pod \"collect-profiles-29401320-jbp7s\" (UID: \"5f2bd1d5-4ceb-4191-9f8f-255f157acecf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401320-jbp7s" Nov 25 14:00:00 crc kubenswrapper[4702]: I1125 14:00:00.272719 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5f2bd1d5-4ceb-4191-9f8f-255f157acecf-secret-volume\") pod \"collect-profiles-29401320-jbp7s\" (UID: \"5f2bd1d5-4ceb-4191-9f8f-255f157acecf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401320-jbp7s" Nov 25 14:00:00 crc kubenswrapper[4702]: I1125 14:00:00.375442 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpbw8\" (UniqueName: \"kubernetes.io/projected/5f2bd1d5-4ceb-4191-9f8f-255f157acecf-kube-api-access-bpbw8\") pod \"collect-profiles-29401320-jbp7s\" (UID: \"5f2bd1d5-4ceb-4191-9f8f-255f157acecf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401320-jbp7s" Nov 25 14:00:00 crc kubenswrapper[4702]: I1125 14:00:00.375605 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5f2bd1d5-4ceb-4191-9f8f-255f157acecf-config-volume\") pod \"collect-profiles-29401320-jbp7s\" (UID: \"5f2bd1d5-4ceb-4191-9f8f-255f157acecf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401320-jbp7s" Nov 25 14:00:00 crc kubenswrapper[4702]: I1125 14:00:00.375670 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5f2bd1d5-4ceb-4191-9f8f-255f157acecf-secret-volume\") pod \"collect-profiles-29401320-jbp7s\" (UID: \"5f2bd1d5-4ceb-4191-9f8f-255f157acecf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401320-jbp7s" Nov 25 14:00:00 crc kubenswrapper[4702]: I1125 14:00:00.376916 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5f2bd1d5-4ceb-4191-9f8f-255f157acecf-config-volume\") pod \"collect-profiles-29401320-jbp7s\" (UID: \"5f2bd1d5-4ceb-4191-9f8f-255f157acecf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401320-jbp7s" Nov 25 14:00:00 crc kubenswrapper[4702]: I1125 14:00:00.400121 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5f2bd1d5-4ceb-4191-9f8f-255f157acecf-secret-volume\") pod \"collect-profiles-29401320-jbp7s\" (UID: \"5f2bd1d5-4ceb-4191-9f8f-255f157acecf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401320-jbp7s" Nov 25 14:00:00 crc kubenswrapper[4702]: I1125 14:00:00.403288 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpbw8\" (UniqueName: \"kubernetes.io/projected/5f2bd1d5-4ceb-4191-9f8f-255f157acecf-kube-api-access-bpbw8\") pod \"collect-profiles-29401320-jbp7s\" (UID: \"5f2bd1d5-4ceb-4191-9f8f-255f157acecf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401320-jbp7s" Nov 25 14:00:00 crc kubenswrapper[4702]: I1125 14:00:00.503932 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401320-jbp7s" Nov 25 14:00:01 crc kubenswrapper[4702]: I1125 14:00:01.118558 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401320-jbp7s"] Nov 25 14:00:01 crc kubenswrapper[4702]: W1125 14:00:01.127478 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f2bd1d5_4ceb_4191_9f8f_255f157acecf.slice/crio-86159c81d8c1cb82241daa5e7dc14cf2455c37680afff1a2ffcd7fb6d79f8d03 WatchSource:0}: Error finding container 86159c81d8c1cb82241daa5e7dc14cf2455c37680afff1a2ffcd7fb6d79f8d03: Status 404 returned error can't find the container with id 86159c81d8c1cb82241daa5e7dc14cf2455c37680afff1a2ffcd7fb6d79f8d03 Nov 25 14:00:01 crc kubenswrapper[4702]: E1125 14:00:01.736983 4702 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f2bd1d5_4ceb_4191_9f8f_255f157acecf.slice/crio-f1d77d86435c5ad0d3fafbddfd95d04ed3170fdd6ea9a8d8d46690bf0207661b.scope\": RecentStats: unable to find data in memory cache]" Nov 25 14:00:01 crc kubenswrapper[4702]: I1125 14:00:01.979778 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 25 14:00:01 crc kubenswrapper[4702]: I1125 14:00:01.981276 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 14:00:01 crc kubenswrapper[4702]: I1125 14:00:01.983211 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"kube-root-ca.crt" Nov 25 14:00:01 crc kubenswrapper[4702]: I1125 14:00:01.983440 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openshift-service-ca.crt" Nov 25 14:00:01 crc kubenswrapper[4702]: I1125 14:00:01.983685 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"galera-openstack-dockercfg-l8s6m" Nov 25 14:00:01 crc kubenswrapper[4702]: I1125 14:00:01.983761 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts" Nov 25 14:00:01 crc kubenswrapper[4702]: I1125 14:00:01.983789 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config-data" Nov 25 14:00:01 crc kubenswrapper[4702]: I1125 14:00:01.985040 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 25 14:00:01 crc kubenswrapper[4702]: I1125 14:00:01.986461 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 14:00:01 crc kubenswrapper[4702]: I1125 14:00:01.990952 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 25 14:00:01 crc kubenswrapper[4702]: I1125 14:00:01.991990 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 14:00:01 crc kubenswrapper[4702]: I1125 14:00:01.995553 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.000574 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.004843 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.088342 4702 generic.go:334] "Generic (PLEG): container finished" podID="5f2bd1d5-4ceb-4191-9f8f-255f157acecf" containerID="f1d77d86435c5ad0d3fafbddfd95d04ed3170fdd6ea9a8d8d46690bf0207661b" exitCode=0 Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.088410 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401320-jbp7s" event={"ID":"5f2bd1d5-4ceb-4191-9f8f-255f157acecf","Type":"ContainerDied","Data":"f1d77d86435c5ad0d3fafbddfd95d04ed3170fdd6ea9a8d8d46690bf0207661b"} Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.088457 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401320-jbp7s" event={"ID":"5f2bd1d5-4ceb-4191-9f8f-255f157acecf","Type":"ContainerStarted","Data":"86159c81d8c1cb82241daa5e7dc14cf2455c37680afff1a2ffcd7fb6d79f8d03"} Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.098705 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f17baa10-42c4-4620-9b3c-a5dde9f78229-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f17baa10-42c4-4620-9b3c-a5dde9f78229\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.098751 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8a586ccc-2c61-418c-a84c-51730c841205-config-data-default\") pod \"openstack-galera-1\" (UID: \"8a586ccc-2c61-418c-a84c-51730c841205\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.098782 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/bbd4f8e2-cb5f-45ff-ad8f-0faef399b018-config-data-default\") pod \"openstack-galera-2\" (UID: \"bbd4f8e2-cb5f-45ff-ad8f-0faef399b018\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.098811 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/bbd4f8e2-cb5f-45ff-ad8f-0faef399b018-config-data-generated\") pod \"openstack-galera-2\" (UID: \"bbd4f8e2-cb5f-45ff-ad8f-0faef399b018\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.098844 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-2\" (UID: \"bbd4f8e2-cb5f-45ff-ad8f-0faef399b018\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.098965 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f17baa10-42c4-4620-9b3c-a5dde9f78229-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f17baa10-42c4-4620-9b3c-a5dde9f78229\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.099078 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbd4f8e2-cb5f-45ff-ad8f-0faef399b018-operator-scripts\") pod \"openstack-galera-2\" (UID: \"bbd4f8e2-cb5f-45ff-ad8f-0faef399b018\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.099131 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a586ccc-2c61-418c-a84c-51730c841205-operator-scripts\") pod \"openstack-galera-1\" (UID: \"8a586ccc-2c61-418c-a84c-51730c841205\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.099339 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srnt8\" (UniqueName: \"kubernetes.io/projected/bbd4f8e2-cb5f-45ff-ad8f-0faef399b018-kube-api-access-srnt8\") pod \"openstack-galera-2\" (UID: \"bbd4f8e2-cb5f-45ff-ad8f-0faef399b018\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.099975 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f17baa10-42c4-4620-9b3c-a5dde9f78229-config-data-default\") pod \"openstack-galera-0\" (UID: \"f17baa10-42c4-4620-9b3c-a5dde9f78229\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.100012 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzqdh\" (UniqueName: \"kubernetes.io/projected/8a586ccc-2c61-418c-a84c-51730c841205-kube-api-access-qzqdh\") pod \"openstack-galera-1\" (UID: \"8a586ccc-2c61-418c-a84c-51730c841205\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.100065 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"f17baa10-42c4-4620-9b3c-a5dde9f78229\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.100200 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f17baa10-42c4-4620-9b3c-a5dde9f78229-kolla-config\") pod \"openstack-galera-0\" (UID: \"f17baa10-42c4-4620-9b3c-a5dde9f78229\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.100257 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-1\" (UID: \"8a586ccc-2c61-418c-a84c-51730c841205\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.100283 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8a586ccc-2c61-418c-a84c-51730c841205-kolla-config\") pod \"openstack-galera-1\" (UID: \"8a586ccc-2c61-418c-a84c-51730c841205\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.100341 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2n7t2\" (UniqueName: \"kubernetes.io/projected/f17baa10-42c4-4620-9b3c-a5dde9f78229-kube-api-access-2n7t2\") pod \"openstack-galera-0\" (UID: \"f17baa10-42c4-4620-9b3c-a5dde9f78229\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.100365 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bbd4f8e2-cb5f-45ff-ad8f-0faef399b018-kolla-config\") pod \"openstack-galera-2\" (UID: \"bbd4f8e2-cb5f-45ff-ad8f-0faef399b018\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.100392 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8a586ccc-2c61-418c-a84c-51730c841205-config-data-generated\") pod \"openstack-galera-1\" (UID: \"8a586ccc-2c61-418c-a84c-51730c841205\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.202009 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f17baa10-42c4-4620-9b3c-a5dde9f78229-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f17baa10-42c4-4620-9b3c-a5dde9f78229\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.202076 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbd4f8e2-cb5f-45ff-ad8f-0faef399b018-operator-scripts\") pod \"openstack-galera-2\" (UID: \"bbd4f8e2-cb5f-45ff-ad8f-0faef399b018\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.202121 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a586ccc-2c61-418c-a84c-51730c841205-operator-scripts\") pod \"openstack-galera-1\" (UID: \"8a586ccc-2c61-418c-a84c-51730c841205\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.202146 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srnt8\" (UniqueName: \"kubernetes.io/projected/bbd4f8e2-cb5f-45ff-ad8f-0faef399b018-kube-api-access-srnt8\") pod \"openstack-galera-2\" (UID: \"bbd4f8e2-cb5f-45ff-ad8f-0faef399b018\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.202169 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f17baa10-42c4-4620-9b3c-a5dde9f78229-config-data-default\") pod \"openstack-galera-0\" (UID: \"f17baa10-42c4-4620-9b3c-a5dde9f78229\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.202204 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzqdh\" (UniqueName: \"kubernetes.io/projected/8a586ccc-2c61-418c-a84c-51730c841205-kube-api-access-qzqdh\") pod \"openstack-galera-1\" (UID: \"8a586ccc-2c61-418c-a84c-51730c841205\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.202228 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"f17baa10-42c4-4620-9b3c-a5dde9f78229\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.202286 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f17baa10-42c4-4620-9b3c-a5dde9f78229-kolla-config\") pod \"openstack-galera-0\" (UID: \"f17baa10-42c4-4620-9b3c-a5dde9f78229\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.202304 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-1\" (UID: \"8a586ccc-2c61-418c-a84c-51730c841205\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.202342 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8a586ccc-2c61-418c-a84c-51730c841205-kolla-config\") pod \"openstack-galera-1\" (UID: \"8a586ccc-2c61-418c-a84c-51730c841205\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.202373 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2n7t2\" (UniqueName: \"kubernetes.io/projected/f17baa10-42c4-4620-9b3c-a5dde9f78229-kube-api-access-2n7t2\") pod \"openstack-galera-0\" (UID: \"f17baa10-42c4-4620-9b3c-a5dde9f78229\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.202390 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bbd4f8e2-cb5f-45ff-ad8f-0faef399b018-kolla-config\") pod \"openstack-galera-2\" (UID: \"bbd4f8e2-cb5f-45ff-ad8f-0faef399b018\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.202426 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8a586ccc-2c61-418c-a84c-51730c841205-config-data-generated\") pod \"openstack-galera-1\" (UID: \"8a586ccc-2c61-418c-a84c-51730c841205\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.202459 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f17baa10-42c4-4620-9b3c-a5dde9f78229-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f17baa10-42c4-4620-9b3c-a5dde9f78229\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.202479 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f17baa10-42c4-4620-9b3c-a5dde9f78229-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f17baa10-42c4-4620-9b3c-a5dde9f78229\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.202509 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8a586ccc-2c61-418c-a84c-51730c841205-config-data-default\") pod \"openstack-galera-1\" (UID: \"8a586ccc-2c61-418c-a84c-51730c841205\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.202530 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/bbd4f8e2-cb5f-45ff-ad8f-0faef399b018-config-data-default\") pod \"openstack-galera-2\" (UID: \"bbd4f8e2-cb5f-45ff-ad8f-0faef399b018\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.202551 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/bbd4f8e2-cb5f-45ff-ad8f-0faef399b018-config-data-generated\") pod \"openstack-galera-2\" (UID: \"bbd4f8e2-cb5f-45ff-ad8f-0faef399b018\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.202623 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-2\" (UID: \"bbd4f8e2-cb5f-45ff-ad8f-0faef399b018\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.203031 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-2\" (UID: \"bbd4f8e2-cb5f-45ff-ad8f-0faef399b018\") device mount path \"/mnt/openstack/pv11\"" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.203187 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f17baa10-42c4-4620-9b3c-a5dde9f78229-kolla-config\") pod \"openstack-galera-0\" (UID: \"f17baa10-42c4-4620-9b3c-a5dde9f78229\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.203644 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-1\" (UID: \"8a586ccc-2c61-418c-a84c-51730c841205\") device mount path \"/mnt/openstack/pv06\"" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.204723 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8a586ccc-2c61-418c-a84c-51730c841205-kolla-config\") pod \"openstack-galera-1\" (UID: \"8a586ccc-2c61-418c-a84c-51730c841205\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.205167 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbd4f8e2-cb5f-45ff-ad8f-0faef399b018-operator-scripts\") pod \"openstack-galera-2\" (UID: \"bbd4f8e2-cb5f-45ff-ad8f-0faef399b018\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.205291 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/bbd4f8e2-cb5f-45ff-ad8f-0faef399b018-config-data-default\") pod \"openstack-galera-2\" (UID: \"bbd4f8e2-cb5f-45ff-ad8f-0faef399b018\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.205454 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"f17baa10-42c4-4620-9b3c-a5dde9f78229\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.205498 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f17baa10-42c4-4620-9b3c-a5dde9f78229-config-data-default\") pod \"openstack-galera-0\" (UID: \"f17baa10-42c4-4620-9b3c-a5dde9f78229\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.205557 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8a586ccc-2c61-418c-a84c-51730c841205-config-data-generated\") pod \"openstack-galera-1\" (UID: \"8a586ccc-2c61-418c-a84c-51730c841205\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.206536 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bbd4f8e2-cb5f-45ff-ad8f-0faef399b018-kolla-config\") pod \"openstack-galera-2\" (UID: \"bbd4f8e2-cb5f-45ff-ad8f-0faef399b018\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.207384 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f17baa10-42c4-4620-9b3c-a5dde9f78229-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f17baa10-42c4-4620-9b3c-a5dde9f78229\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.207413 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8a586ccc-2c61-418c-a84c-51730c841205-config-data-default\") pod \"openstack-galera-1\" (UID: \"8a586ccc-2c61-418c-a84c-51730c841205\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.208167 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a586ccc-2c61-418c-a84c-51730c841205-operator-scripts\") pod \"openstack-galera-1\" (UID: \"8a586ccc-2c61-418c-a84c-51730c841205\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.216810 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/bbd4f8e2-cb5f-45ff-ad8f-0faef399b018-config-data-generated\") pod \"openstack-galera-2\" (UID: \"bbd4f8e2-cb5f-45ff-ad8f-0faef399b018\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.224930 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-2\" (UID: \"bbd4f8e2-cb5f-45ff-ad8f-0faef399b018\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.226618 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2n7t2\" (UniqueName: \"kubernetes.io/projected/f17baa10-42c4-4620-9b3c-a5dde9f78229-kube-api-access-2n7t2\") pod \"openstack-galera-0\" (UID: \"f17baa10-42c4-4620-9b3c-a5dde9f78229\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.226971 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzqdh\" (UniqueName: \"kubernetes.io/projected/8a586ccc-2c61-418c-a84c-51730c841205-kube-api-access-qzqdh\") pod \"openstack-galera-1\" (UID: \"8a586ccc-2c61-418c-a84c-51730c841205\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.228289 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"f17baa10-42c4-4620-9b3c-a5dde9f78229\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.229656 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srnt8\" (UniqueName: \"kubernetes.io/projected/bbd4f8e2-cb5f-45ff-ad8f-0faef399b018-kube-api-access-srnt8\") pod \"openstack-galera-2\" (UID: \"bbd4f8e2-cb5f-45ff-ad8f-0faef399b018\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.233385 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-1\" (UID: \"8a586ccc-2c61-418c-a84c-51730c841205\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.305193 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.320994 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.344072 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.796652 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 25 14:00:02 crc kubenswrapper[4702]: I1125 14:00:02.919943 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 25 14:00:02 crc kubenswrapper[4702]: W1125 14:00:02.923421 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a586ccc_2c61_418c_a84c_51730c841205.slice/crio-2c15ca8abedb878a46db58b0d94234510a5070c366db767ff52dec7c14935909 WatchSource:0}: Error finding container 2c15ca8abedb878a46db58b0d94234510a5070c366db767ff52dec7c14935909: Status 404 returned error can't find the container with id 2c15ca8abedb878a46db58b0d94234510a5070c366db767ff52dec7c14935909 Nov 25 14:00:03 crc kubenswrapper[4702]: I1125 14:00:03.078645 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 25 14:00:03 crc kubenswrapper[4702]: W1125 14:00:03.086681 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf17baa10_42c4_4620_9b3c_a5dde9f78229.slice/crio-b69346b9348c03f1d3d0948ad4fca421a9e7d25f760dca47c81ac61980053b0a WatchSource:0}: Error finding container b69346b9348c03f1d3d0948ad4fca421a9e7d25f760dca47c81ac61980053b0a: Status 404 returned error can't find the container with id b69346b9348c03f1d3d0948ad4fca421a9e7d25f760dca47c81ac61980053b0a Nov 25 14:00:03 crc kubenswrapper[4702]: I1125 14:00:03.097039 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"f17baa10-42c4-4620-9b3c-a5dde9f78229","Type":"ContainerStarted","Data":"b69346b9348c03f1d3d0948ad4fca421a9e7d25f760dca47c81ac61980053b0a"} Nov 25 14:00:03 crc kubenswrapper[4702]: I1125 14:00:03.098178 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"bbd4f8e2-cb5f-45ff-ad8f-0faef399b018","Type":"ContainerStarted","Data":"91e34d070f1171e294715ecf9f72bca7337c6ad8083667531d4d996c49353f9e"} Nov 25 14:00:03 crc kubenswrapper[4702]: I1125 14:00:03.099132 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"8a586ccc-2c61-418c-a84c-51730c841205","Type":"ContainerStarted","Data":"2c15ca8abedb878a46db58b0d94234510a5070c366db767ff52dec7c14935909"} Nov 25 14:00:03 crc kubenswrapper[4702]: I1125 14:00:03.365432 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401320-jbp7s" Nov 25 14:00:03 crc kubenswrapper[4702]: I1125 14:00:03.424704 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5f2bd1d5-4ceb-4191-9f8f-255f157acecf-secret-volume\") pod \"5f2bd1d5-4ceb-4191-9f8f-255f157acecf\" (UID: \"5f2bd1d5-4ceb-4191-9f8f-255f157acecf\") " Nov 25 14:00:03 crc kubenswrapper[4702]: I1125 14:00:03.424808 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5f2bd1d5-4ceb-4191-9f8f-255f157acecf-config-volume\") pod \"5f2bd1d5-4ceb-4191-9f8f-255f157acecf\" (UID: \"5f2bd1d5-4ceb-4191-9f8f-255f157acecf\") " Nov 25 14:00:03 crc kubenswrapper[4702]: I1125 14:00:03.424851 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpbw8\" (UniqueName: \"kubernetes.io/projected/5f2bd1d5-4ceb-4191-9f8f-255f157acecf-kube-api-access-bpbw8\") pod \"5f2bd1d5-4ceb-4191-9f8f-255f157acecf\" (UID: \"5f2bd1d5-4ceb-4191-9f8f-255f157acecf\") " Nov 25 14:00:03 crc kubenswrapper[4702]: I1125 14:00:03.426440 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f2bd1d5-4ceb-4191-9f8f-255f157acecf-config-volume" (OuterVolumeSpecName: "config-volume") pod "5f2bd1d5-4ceb-4191-9f8f-255f157acecf" (UID: "5f2bd1d5-4ceb-4191-9f8f-255f157acecf"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 14:00:03 crc kubenswrapper[4702]: I1125 14:00:03.439084 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f2bd1d5-4ceb-4191-9f8f-255f157acecf-kube-api-access-bpbw8" (OuterVolumeSpecName: "kube-api-access-bpbw8") pod "5f2bd1d5-4ceb-4191-9f8f-255f157acecf" (UID: "5f2bd1d5-4ceb-4191-9f8f-255f157acecf"). InnerVolumeSpecName "kube-api-access-bpbw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:00:03 crc kubenswrapper[4702]: I1125 14:00:03.439795 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f2bd1d5-4ceb-4191-9f8f-255f157acecf-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5f2bd1d5-4ceb-4191-9f8f-255f157acecf" (UID: "5f2bd1d5-4ceb-4191-9f8f-255f157acecf"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:00:03 crc kubenswrapper[4702]: I1125 14:00:03.526798 4702 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5f2bd1d5-4ceb-4191-9f8f-255f157acecf-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 14:00:03 crc kubenswrapper[4702]: I1125 14:00:03.526860 4702 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5f2bd1d5-4ceb-4191-9f8f-255f157acecf-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 14:00:03 crc kubenswrapper[4702]: I1125 14:00:03.526879 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpbw8\" (UniqueName: \"kubernetes.io/projected/5f2bd1d5-4ceb-4191-9f8f-255f157acecf-kube-api-access-bpbw8\") on node \"crc\" DevicePath \"\"" Nov 25 14:00:03 crc kubenswrapper[4702]: I1125 14:00:03.962696 4702 patch_prober.go:28] interesting pod/machine-config-daemon-qqr76 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 14:00:03 crc kubenswrapper[4702]: I1125 14:00:03.962780 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 14:00:04 crc kubenswrapper[4702]: I1125 14:00:04.112006 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401320-jbp7s" event={"ID":"5f2bd1d5-4ceb-4191-9f8f-255f157acecf","Type":"ContainerDied","Data":"86159c81d8c1cb82241daa5e7dc14cf2455c37680afff1a2ffcd7fb6d79f8d03"} Nov 25 14:00:04 crc kubenswrapper[4702]: I1125 14:00:04.112066 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="86159c81d8c1cb82241daa5e7dc14cf2455c37680afff1a2ffcd7fb6d79f8d03" Nov 25 14:00:04 crc kubenswrapper[4702]: I1125 14:00:04.112088 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401320-jbp7s" Nov 25 14:00:24 crc kubenswrapper[4702]: E1125 14:00:24.712571 4702 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb@sha256:10452e2144368e2f128c8fb8ef9e54880b06ef1d71d9f084a0217dcb099c51ce" Nov 25 14:00:24 crc kubenswrapper[4702]: E1125 14:00:24.713356 4702 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:10452e2144368e2f128c8fb8ef9e54880b06ef1d71d9f084a0217dcb099c51ce,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qzqdh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-1_glance-kuttl-tests(8a586ccc-2c61-418c-a84c-51730c841205): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 14:00:24 crc kubenswrapper[4702]: E1125 14:00:24.714552 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="glance-kuttl-tests/openstack-galera-1" podUID="8a586ccc-2c61-418c-a84c-51730c841205" Nov 25 14:00:25 crc kubenswrapper[4702]: E1125 14:00:25.248701 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb@sha256:10452e2144368e2f128c8fb8ef9e54880b06ef1d71d9f084a0217dcb099c51ce\\\"\"" pod="glance-kuttl-tests/openstack-galera-1" podUID="8a586ccc-2c61-418c-a84c-51730c841205" Nov 25 14:00:25 crc kubenswrapper[4702]: E1125 14:00:25.258330 4702 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy@sha256:d28df2924a366ed857d6c2c14baac9741238032d41f3d02c12cd757189b68b8a" Nov 25 14:00:25 crc kubenswrapper[4702]: E1125 14:00:25.258483 4702 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy@sha256:d28df2924a366ed857d6c2c14baac9741238032d41f3d02c12cd757189b68b8a,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:infra-operator.v0.0.1,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:apiservice-cert,ReadOnly:false,MountPath:/apiserver.local.config/certificates,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jzjbj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-79898f8746-rfpp7_openstack-operators(e13b9e67-df17-4851-a3c5-5037b994637d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 14:00:25 crc kubenswrapper[4702]: E1125 14:00:25.259737 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/infra-operator-controller-manager-79898f8746-rfpp7" podUID="e13b9e67-df17-4851-a3c5-5037b994637d" Nov 25 14:00:25 crc kubenswrapper[4702]: E1125 14:00:25.544533 4702 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb@sha256:10452e2144368e2f128c8fb8ef9e54880b06ef1d71d9f084a0217dcb099c51ce" Nov 25 14:00:25 crc kubenswrapper[4702]: E1125 14:00:25.544696 4702 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:10452e2144368e2f128c8fb8ef9e54880b06ef1d71d9f084a0217dcb099c51ce,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2n7t2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_glance-kuttl-tests(f17baa10-42c4-4620-9b3c-a5dde9f78229): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 14:00:25 crc kubenswrapper[4702]: E1125 14:00:25.545870 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="glance-kuttl-tests/openstack-galera-0" podUID="f17baa10-42c4-4620-9b3c-a5dde9f78229" Nov 25 14:00:25 crc kubenswrapper[4702]: E1125 14:00:25.704909 4702 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb@sha256:10452e2144368e2f128c8fb8ef9e54880b06ef1d71d9f084a0217dcb099c51ce" Nov 25 14:00:25 crc kubenswrapper[4702]: E1125 14:00:25.705070 4702 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:10452e2144368e2f128c8fb8ef9e54880b06ef1d71d9f084a0217dcb099c51ce,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-srnt8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-2_glance-kuttl-tests(bbd4f8e2-cb5f-45ff-ad8f-0faef399b018): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 14:00:25 crc kubenswrapper[4702]: E1125 14:00:25.706526 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="glance-kuttl-tests/openstack-galera-2" podUID="bbd4f8e2-cb5f-45ff-ad8f-0faef399b018" Nov 25 14:00:26 crc kubenswrapper[4702]: E1125 14:00:26.253579 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb@sha256:10452e2144368e2f128c8fb8ef9e54880b06ef1d71d9f084a0217dcb099c51ce\\\"\"" pod="glance-kuttl-tests/openstack-galera-2" podUID="bbd4f8e2-cb5f-45ff-ad8f-0faef399b018" Nov 25 14:00:26 crc kubenswrapper[4702]: E1125 14:00:26.253598 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb@sha256:10452e2144368e2f128c8fb8ef9e54880b06ef1d71d9f084a0217dcb099c51ce\\\"\"" pod="glance-kuttl-tests/openstack-galera-0" podUID="f17baa10-42c4-4620-9b3c-a5dde9f78229" Nov 25 14:00:26 crc kubenswrapper[4702]: E1125 14:00:26.253959 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy@sha256:d28df2924a366ed857d6c2c14baac9741238032d41f3d02c12cd757189b68b8a\\\"\"" pod="openstack-operators/infra-operator-controller-manager-79898f8746-rfpp7" podUID="e13b9e67-df17-4851-a3c5-5037b994637d" Nov 25 14:00:33 crc kubenswrapper[4702]: I1125 14:00:33.963003 4702 patch_prober.go:28] interesting pod/machine-config-daemon-qqr76 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 14:00:33 crc kubenswrapper[4702]: I1125 14:00:33.963993 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 14:00:33 crc kubenswrapper[4702]: I1125 14:00:33.964082 4702 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" Nov 25 14:00:33 crc kubenswrapper[4702]: I1125 14:00:33.965025 4702 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"63ddfd96ea126ce49009797997159362938b731d35159663d233bd6f5c4e30d7"} pod="openshift-machine-config-operator/machine-config-daemon-qqr76" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 14:00:33 crc kubenswrapper[4702]: I1125 14:00:33.965094 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" containerID="cri-o://63ddfd96ea126ce49009797997159362938b731d35159663d233bd6f5c4e30d7" gracePeriod=600 Nov 25 14:00:34 crc kubenswrapper[4702]: I1125 14:00:34.316226 4702 generic.go:334] "Generic (PLEG): container finished" podID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerID="63ddfd96ea126ce49009797997159362938b731d35159663d233bd6f5c4e30d7" exitCode=0 Nov 25 14:00:34 crc kubenswrapper[4702]: I1125 14:00:34.316300 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" event={"ID":"3c491818-b0bb-4d82-a031-96e2dfac8c27","Type":"ContainerDied","Data":"63ddfd96ea126ce49009797997159362938b731d35159663d233bd6f5c4e30d7"} Nov 25 14:00:34 crc kubenswrapper[4702]: I1125 14:00:34.317165 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" event={"ID":"3c491818-b0bb-4d82-a031-96e2dfac8c27","Type":"ContainerStarted","Data":"03504dcb763332c521878c39bb07b19a232e8fc711f4c606ea5726c5c5f4da1a"} Nov 25 14:00:34 crc kubenswrapper[4702]: I1125 14:00:34.317230 4702 scope.go:117] "RemoveContainer" containerID="a1697a0482710a529c531a772ee7967b9b00f72bbd70b22b58a644272dc2e190" Nov 25 14:00:35 crc kubenswrapper[4702]: I1125 14:00:35.381350 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-79898f8746-rfpp7" Nov 25 14:00:35 crc kubenswrapper[4702]: I1125 14:00:35.388793 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-79898f8746-rfpp7" Nov 25 14:00:36 crc kubenswrapper[4702]: I1125 14:00:36.196723 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-sjhwl"] Nov 25 14:00:36 crc kubenswrapper[4702]: E1125 14:00:36.197426 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f2bd1d5-4ceb-4191-9f8f-255f157acecf" containerName="collect-profiles" Nov 25 14:00:36 crc kubenswrapper[4702]: I1125 14:00:36.197441 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f2bd1d5-4ceb-4191-9f8f-255f157acecf" containerName="collect-profiles" Nov 25 14:00:36 crc kubenswrapper[4702]: I1125 14:00:36.197552 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f2bd1d5-4ceb-4191-9f8f-255f157acecf" containerName="collect-profiles" Nov 25 14:00:36 crc kubenswrapper[4702]: I1125 14:00:36.198348 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sjhwl" Nov 25 14:00:36 crc kubenswrapper[4702]: I1125 14:00:36.213178 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sjhwl"] Nov 25 14:00:36 crc kubenswrapper[4702]: I1125 14:00:36.373954 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/147df53b-0c7b-4de1-b9b5-ee311584fdc7-utilities\") pod \"certified-operators-sjhwl\" (UID: \"147df53b-0c7b-4de1-b9b5-ee311584fdc7\") " pod="openshift-marketplace/certified-operators-sjhwl" Nov 25 14:00:36 crc kubenswrapper[4702]: I1125 14:00:36.374090 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/147df53b-0c7b-4de1-b9b5-ee311584fdc7-catalog-content\") pod \"certified-operators-sjhwl\" (UID: \"147df53b-0c7b-4de1-b9b5-ee311584fdc7\") " pod="openshift-marketplace/certified-operators-sjhwl" Nov 25 14:00:36 crc kubenswrapper[4702]: I1125 14:00:36.374186 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlvv9\" (UniqueName: \"kubernetes.io/projected/147df53b-0c7b-4de1-b9b5-ee311584fdc7-kube-api-access-vlvv9\") pod \"certified-operators-sjhwl\" (UID: \"147df53b-0c7b-4de1-b9b5-ee311584fdc7\") " pod="openshift-marketplace/certified-operators-sjhwl" Nov 25 14:00:36 crc kubenswrapper[4702]: I1125 14:00:36.475553 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/147df53b-0c7b-4de1-b9b5-ee311584fdc7-catalog-content\") pod \"certified-operators-sjhwl\" (UID: \"147df53b-0c7b-4de1-b9b5-ee311584fdc7\") " pod="openshift-marketplace/certified-operators-sjhwl" Nov 25 14:00:36 crc kubenswrapper[4702]: I1125 14:00:36.475657 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlvv9\" (UniqueName: \"kubernetes.io/projected/147df53b-0c7b-4de1-b9b5-ee311584fdc7-kube-api-access-vlvv9\") pod \"certified-operators-sjhwl\" (UID: \"147df53b-0c7b-4de1-b9b5-ee311584fdc7\") " pod="openshift-marketplace/certified-operators-sjhwl" Nov 25 14:00:36 crc kubenswrapper[4702]: I1125 14:00:36.475706 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/147df53b-0c7b-4de1-b9b5-ee311584fdc7-utilities\") pod \"certified-operators-sjhwl\" (UID: \"147df53b-0c7b-4de1-b9b5-ee311584fdc7\") " pod="openshift-marketplace/certified-operators-sjhwl" Nov 25 14:00:36 crc kubenswrapper[4702]: I1125 14:00:36.476176 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/147df53b-0c7b-4de1-b9b5-ee311584fdc7-catalog-content\") pod \"certified-operators-sjhwl\" (UID: \"147df53b-0c7b-4de1-b9b5-ee311584fdc7\") " pod="openshift-marketplace/certified-operators-sjhwl" Nov 25 14:00:36 crc kubenswrapper[4702]: I1125 14:00:36.476274 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/147df53b-0c7b-4de1-b9b5-ee311584fdc7-utilities\") pod \"certified-operators-sjhwl\" (UID: \"147df53b-0c7b-4de1-b9b5-ee311584fdc7\") " pod="openshift-marketplace/certified-operators-sjhwl" Nov 25 14:00:36 crc kubenswrapper[4702]: I1125 14:00:36.500136 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlvv9\" (UniqueName: \"kubernetes.io/projected/147df53b-0c7b-4de1-b9b5-ee311584fdc7-kube-api-access-vlvv9\") pod \"certified-operators-sjhwl\" (UID: \"147df53b-0c7b-4de1-b9b5-ee311584fdc7\") " pod="openshift-marketplace/certified-operators-sjhwl" Nov 25 14:00:36 crc kubenswrapper[4702]: I1125 14:00:36.516087 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sjhwl" Nov 25 14:00:37 crc kubenswrapper[4702]: W1125 14:00:37.835403 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod147df53b_0c7b_4de1_b9b5_ee311584fdc7.slice/crio-de608e7d4dc0df04d87203bf82b5060d81878d3ab5fcc9ff5e3f979a01de8055 WatchSource:0}: Error finding container de608e7d4dc0df04d87203bf82b5060d81878d3ab5fcc9ff5e3f979a01de8055: Status 404 returned error can't find the container with id de608e7d4dc0df04d87203bf82b5060d81878d3ab5fcc9ff5e3f979a01de8055 Nov 25 14:00:37 crc kubenswrapper[4702]: I1125 14:00:37.837224 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sjhwl"] Nov 25 14:00:38 crc kubenswrapper[4702]: I1125 14:00:38.361345 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79898f8746-rfpp7" event={"ID":"e13b9e67-df17-4851-a3c5-5037b994637d","Type":"ContainerStarted","Data":"6becdd46f9deabede59617e57e17d9847efe45b420bbe149a85b3e327ea0c1e1"} Nov 25 14:00:38 crc kubenswrapper[4702]: I1125 14:00:38.364869 4702 generic.go:334] "Generic (PLEG): container finished" podID="147df53b-0c7b-4de1-b9b5-ee311584fdc7" containerID="068bd2be665b609399bac7c680646827506b632975549c74e0d791366d3402e1" exitCode=0 Nov 25 14:00:38 crc kubenswrapper[4702]: I1125 14:00:38.364960 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjhwl" event={"ID":"147df53b-0c7b-4de1-b9b5-ee311584fdc7","Type":"ContainerDied","Data":"068bd2be665b609399bac7c680646827506b632975549c74e0d791366d3402e1"} Nov 25 14:00:38 crc kubenswrapper[4702]: I1125 14:00:38.365285 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjhwl" event={"ID":"147df53b-0c7b-4de1-b9b5-ee311584fdc7","Type":"ContainerStarted","Data":"de608e7d4dc0df04d87203bf82b5060d81878d3ab5fcc9ff5e3f979a01de8055"} Nov 25 14:00:38 crc kubenswrapper[4702]: I1125 14:00:38.391105 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-79898f8746-rfpp7" podStartSLOduration=1.796680291 podStartE2EDuration="43.391076573s" podCreationTimestamp="2025-11-25 13:59:55 +0000 UTC" firstStartedPulling="2025-11-25 13:59:55.810183206 +0000 UTC m=+804.972833656" lastFinishedPulling="2025-11-25 14:00:37.404579488 +0000 UTC m=+846.567229938" observedRunningTime="2025-11-25 14:00:38.388272942 +0000 UTC m=+847.550923432" watchObservedRunningTime="2025-11-25 14:00:38.391076573 +0000 UTC m=+847.553727033" Nov 25 14:00:39 crc kubenswrapper[4702]: I1125 14:00:39.380429 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjhwl" event={"ID":"147df53b-0c7b-4de1-b9b5-ee311584fdc7","Type":"ContainerStarted","Data":"75c4fd157d2d9ca8bf4819d111e99da9721f9123d040914affabb33c51bff4b7"} Nov 25 14:00:39 crc kubenswrapper[4702]: I1125 14:00:39.906276 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 25 14:00:39 crc kubenswrapper[4702]: I1125 14:00:39.907902 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Nov 25 14:00:39 crc kubenswrapper[4702]: I1125 14:00:39.910366 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"memcached-config-data" Nov 25 14:00:39 crc kubenswrapper[4702]: I1125 14:00:39.911585 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"memcached-memcached-dockercfg-f67rp" Nov 25 14:00:39 crc kubenswrapper[4702]: I1125 14:00:39.934795 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 25 14:00:40 crc kubenswrapper[4702]: I1125 14:00:40.035989 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ac997ffc-32f2-4065-bc24-9982a8e306b6-kolla-config\") pod \"memcached-0\" (UID: \"ac997ffc-32f2-4065-bc24-9982a8e306b6\") " pod="glance-kuttl-tests/memcached-0" Nov 25 14:00:40 crc kubenswrapper[4702]: I1125 14:00:40.036047 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7k95k\" (UniqueName: \"kubernetes.io/projected/ac997ffc-32f2-4065-bc24-9982a8e306b6-kube-api-access-7k95k\") pod \"memcached-0\" (UID: \"ac997ffc-32f2-4065-bc24-9982a8e306b6\") " pod="glance-kuttl-tests/memcached-0" Nov 25 14:00:40 crc kubenswrapper[4702]: I1125 14:00:40.036252 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ac997ffc-32f2-4065-bc24-9982a8e306b6-config-data\") pod \"memcached-0\" (UID: \"ac997ffc-32f2-4065-bc24-9982a8e306b6\") " pod="glance-kuttl-tests/memcached-0" Nov 25 14:00:40 crc kubenswrapper[4702]: I1125 14:00:40.138129 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ac997ffc-32f2-4065-bc24-9982a8e306b6-kolla-config\") pod \"memcached-0\" (UID: \"ac997ffc-32f2-4065-bc24-9982a8e306b6\") " pod="glance-kuttl-tests/memcached-0" Nov 25 14:00:40 crc kubenswrapper[4702]: I1125 14:00:40.138193 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7k95k\" (UniqueName: \"kubernetes.io/projected/ac997ffc-32f2-4065-bc24-9982a8e306b6-kube-api-access-7k95k\") pod \"memcached-0\" (UID: \"ac997ffc-32f2-4065-bc24-9982a8e306b6\") " pod="glance-kuttl-tests/memcached-0" Nov 25 14:00:40 crc kubenswrapper[4702]: I1125 14:00:40.138340 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ac997ffc-32f2-4065-bc24-9982a8e306b6-config-data\") pod \"memcached-0\" (UID: \"ac997ffc-32f2-4065-bc24-9982a8e306b6\") " pod="glance-kuttl-tests/memcached-0" Nov 25 14:00:40 crc kubenswrapper[4702]: I1125 14:00:40.139468 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ac997ffc-32f2-4065-bc24-9982a8e306b6-config-data\") pod \"memcached-0\" (UID: \"ac997ffc-32f2-4065-bc24-9982a8e306b6\") " pod="glance-kuttl-tests/memcached-0" Nov 25 14:00:40 crc kubenswrapper[4702]: I1125 14:00:40.139472 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ac997ffc-32f2-4065-bc24-9982a8e306b6-kolla-config\") pod \"memcached-0\" (UID: \"ac997ffc-32f2-4065-bc24-9982a8e306b6\") " pod="glance-kuttl-tests/memcached-0" Nov 25 14:00:40 crc kubenswrapper[4702]: I1125 14:00:40.164311 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7k95k\" (UniqueName: \"kubernetes.io/projected/ac997ffc-32f2-4065-bc24-9982a8e306b6-kube-api-access-7k95k\") pod \"memcached-0\" (UID: \"ac997ffc-32f2-4065-bc24-9982a8e306b6\") " pod="glance-kuttl-tests/memcached-0" Nov 25 14:00:40 crc kubenswrapper[4702]: I1125 14:00:40.227472 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Nov 25 14:00:40 crc kubenswrapper[4702]: I1125 14:00:40.402439 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"f17baa10-42c4-4620-9b3c-a5dde9f78229","Type":"ContainerStarted","Data":"ac0330bdea5f28246e7216e4037f5e5ae1cb74c08cdc8de157d30c4676a16787"} Nov 25 14:00:40 crc kubenswrapper[4702]: I1125 14:00:40.418017 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"bbd4f8e2-cb5f-45ff-ad8f-0faef399b018","Type":"ContainerStarted","Data":"4961265c62b5dc7b09f0fd529be6f7316c94e279538514c8bbffd00a07d6ab1d"} Nov 25 14:00:40 crc kubenswrapper[4702]: I1125 14:00:40.440925 4702 generic.go:334] "Generic (PLEG): container finished" podID="147df53b-0c7b-4de1-b9b5-ee311584fdc7" containerID="75c4fd157d2d9ca8bf4819d111e99da9721f9123d040914affabb33c51bff4b7" exitCode=0 Nov 25 14:00:40 crc kubenswrapper[4702]: I1125 14:00:40.440981 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjhwl" event={"ID":"147df53b-0c7b-4de1-b9b5-ee311584fdc7","Type":"ContainerDied","Data":"75c4fd157d2d9ca8bf4819d111e99da9721f9123d040914affabb33c51bff4b7"} Nov 25 14:00:40 crc kubenswrapper[4702]: I1125 14:00:40.598462 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 25 14:00:41 crc kubenswrapper[4702]: I1125 14:00:41.451454 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjhwl" event={"ID":"147df53b-0c7b-4de1-b9b5-ee311584fdc7","Type":"ContainerStarted","Data":"69a14209d4fd25c41354e29adc203c641b675f84c612be6619177855e9dc0ea8"} Nov 25 14:00:41 crc kubenswrapper[4702]: I1125 14:00:41.453291 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"8a586ccc-2c61-418c-a84c-51730c841205","Type":"ContainerStarted","Data":"608e3ff4bc93ddc6135275259aa92860868a9760b4ecfff2166ec4b862daafdf"} Nov 25 14:00:41 crc kubenswrapper[4702]: I1125 14:00:41.456178 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"ac997ffc-32f2-4065-bc24-9982a8e306b6","Type":"ContainerStarted","Data":"ed3ca9c1528bec8fc63afd702859e8969d3909ac46cb8d220379821cfd7827cd"} Nov 25 14:00:41 crc kubenswrapper[4702]: I1125 14:00:41.496829 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-sjhwl" podStartSLOduration=2.661898729 podStartE2EDuration="5.49679854s" podCreationTimestamp="2025-11-25 14:00:36 +0000 UTC" firstStartedPulling="2025-11-25 14:00:38.367706338 +0000 UTC m=+847.530356798" lastFinishedPulling="2025-11-25 14:00:41.202606149 +0000 UTC m=+850.365256609" observedRunningTime="2025-11-25 14:00:41.486730258 +0000 UTC m=+850.649380708" watchObservedRunningTime="2025-11-25 14:00:41.49679854 +0000 UTC m=+850.659448990" Nov 25 14:00:43 crc kubenswrapper[4702]: I1125 14:00:43.790735 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-9wl6w"] Nov 25 14:00:43 crc kubenswrapper[4702]: I1125 14:00:43.791784 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-9wl6w" Nov 25 14:00:43 crc kubenswrapper[4702]: I1125 14:00:43.795813 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-index-dockercfg-tw7ng" Nov 25 14:00:43 crc kubenswrapper[4702]: I1125 14:00:43.809898 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-9wl6w"] Nov 25 14:00:43 crc kubenswrapper[4702]: I1125 14:00:43.898482 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bdbs\" (UniqueName: \"kubernetes.io/projected/fc79fae5-b7b4-43b6-89ad-90008bf8dd57-kube-api-access-2bdbs\") pod \"rabbitmq-cluster-operator-index-9wl6w\" (UID: \"fc79fae5-b7b4-43b6-89ad-90008bf8dd57\") " pod="openstack-operators/rabbitmq-cluster-operator-index-9wl6w" Nov 25 14:00:44 crc kubenswrapper[4702]: I1125 14:00:44.000955 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bdbs\" (UniqueName: \"kubernetes.io/projected/fc79fae5-b7b4-43b6-89ad-90008bf8dd57-kube-api-access-2bdbs\") pod \"rabbitmq-cluster-operator-index-9wl6w\" (UID: \"fc79fae5-b7b4-43b6-89ad-90008bf8dd57\") " pod="openstack-operators/rabbitmq-cluster-operator-index-9wl6w" Nov 25 14:00:44 crc kubenswrapper[4702]: I1125 14:00:44.021742 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bdbs\" (UniqueName: \"kubernetes.io/projected/fc79fae5-b7b4-43b6-89ad-90008bf8dd57-kube-api-access-2bdbs\") pod \"rabbitmq-cluster-operator-index-9wl6w\" (UID: \"fc79fae5-b7b4-43b6-89ad-90008bf8dd57\") " pod="openstack-operators/rabbitmq-cluster-operator-index-9wl6w" Nov 25 14:00:44 crc kubenswrapper[4702]: I1125 14:00:44.107892 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-9wl6w" Nov 25 14:00:45 crc kubenswrapper[4702]: I1125 14:00:45.845278 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-9wl6w"] Nov 25 14:00:45 crc kubenswrapper[4702]: W1125 14:00:45.856564 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc79fae5_b7b4_43b6_89ad_90008bf8dd57.slice/crio-febe228e664f5b500c2ccf41ba4ed07bd49100a5a4370ac319662ae9cd43015d WatchSource:0}: Error finding container febe228e664f5b500c2ccf41ba4ed07bd49100a5a4370ac319662ae9cd43015d: Status 404 returned error can't find the container with id febe228e664f5b500c2ccf41ba4ed07bd49100a5a4370ac319662ae9cd43015d Nov 25 14:00:46 crc kubenswrapper[4702]: I1125 14:00:46.503466 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-9wl6w" event={"ID":"fc79fae5-b7b4-43b6-89ad-90008bf8dd57","Type":"ContainerStarted","Data":"febe228e664f5b500c2ccf41ba4ed07bd49100a5a4370ac319662ae9cd43015d"} Nov 25 14:00:46 crc kubenswrapper[4702]: I1125 14:00:46.516975 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-sjhwl" Nov 25 14:00:46 crc kubenswrapper[4702]: I1125 14:00:46.518393 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-sjhwl" Nov 25 14:00:46 crc kubenswrapper[4702]: I1125 14:00:46.573166 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-sjhwl" Nov 25 14:00:47 crc kubenswrapper[4702]: I1125 14:00:47.557915 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-sjhwl" Nov 25 14:00:48 crc kubenswrapper[4702]: I1125 14:00:48.794609 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-r7t6l"] Nov 25 14:00:48 crc kubenswrapper[4702]: I1125 14:00:48.796083 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r7t6l" Nov 25 14:00:48 crc kubenswrapper[4702]: I1125 14:00:48.809811 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r7t6l"] Nov 25 14:00:48 crc kubenswrapper[4702]: I1125 14:00:48.876856 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbl8t\" (UniqueName: \"kubernetes.io/projected/00f64e96-de2b-457c-a365-4dd090524959-kube-api-access-cbl8t\") pod \"community-operators-r7t6l\" (UID: \"00f64e96-de2b-457c-a365-4dd090524959\") " pod="openshift-marketplace/community-operators-r7t6l" Nov 25 14:00:48 crc kubenswrapper[4702]: I1125 14:00:48.876977 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00f64e96-de2b-457c-a365-4dd090524959-catalog-content\") pod \"community-operators-r7t6l\" (UID: \"00f64e96-de2b-457c-a365-4dd090524959\") " pod="openshift-marketplace/community-operators-r7t6l" Nov 25 14:00:48 crc kubenswrapper[4702]: I1125 14:00:48.877063 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00f64e96-de2b-457c-a365-4dd090524959-utilities\") pod \"community-operators-r7t6l\" (UID: \"00f64e96-de2b-457c-a365-4dd090524959\") " pod="openshift-marketplace/community-operators-r7t6l" Nov 25 14:00:48 crc kubenswrapper[4702]: I1125 14:00:48.979263 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbl8t\" (UniqueName: \"kubernetes.io/projected/00f64e96-de2b-457c-a365-4dd090524959-kube-api-access-cbl8t\") pod \"community-operators-r7t6l\" (UID: \"00f64e96-de2b-457c-a365-4dd090524959\") " pod="openshift-marketplace/community-operators-r7t6l" Nov 25 14:00:48 crc kubenswrapper[4702]: I1125 14:00:48.979360 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00f64e96-de2b-457c-a365-4dd090524959-catalog-content\") pod \"community-operators-r7t6l\" (UID: \"00f64e96-de2b-457c-a365-4dd090524959\") " pod="openshift-marketplace/community-operators-r7t6l" Nov 25 14:00:48 crc kubenswrapper[4702]: I1125 14:00:48.979416 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00f64e96-de2b-457c-a365-4dd090524959-utilities\") pod \"community-operators-r7t6l\" (UID: \"00f64e96-de2b-457c-a365-4dd090524959\") " pod="openshift-marketplace/community-operators-r7t6l" Nov 25 14:00:48 crc kubenswrapper[4702]: I1125 14:00:48.979986 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00f64e96-de2b-457c-a365-4dd090524959-utilities\") pod \"community-operators-r7t6l\" (UID: \"00f64e96-de2b-457c-a365-4dd090524959\") " pod="openshift-marketplace/community-operators-r7t6l" Nov 25 14:00:48 crc kubenswrapper[4702]: I1125 14:00:48.980713 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00f64e96-de2b-457c-a365-4dd090524959-catalog-content\") pod \"community-operators-r7t6l\" (UID: \"00f64e96-de2b-457c-a365-4dd090524959\") " pod="openshift-marketplace/community-operators-r7t6l" Nov 25 14:00:49 crc kubenswrapper[4702]: I1125 14:00:49.004252 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbl8t\" (UniqueName: \"kubernetes.io/projected/00f64e96-de2b-457c-a365-4dd090524959-kube-api-access-cbl8t\") pod \"community-operators-r7t6l\" (UID: \"00f64e96-de2b-457c-a365-4dd090524959\") " pod="openshift-marketplace/community-operators-r7t6l" Nov 25 14:00:49 crc kubenswrapper[4702]: I1125 14:00:49.122020 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r7t6l" Nov 25 14:00:49 crc kubenswrapper[4702]: I1125 14:00:49.359748 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r7t6l"] Nov 25 14:00:49 crc kubenswrapper[4702]: I1125 14:00:49.522521 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r7t6l" event={"ID":"00f64e96-de2b-457c-a365-4dd090524959","Type":"ContainerStarted","Data":"534685b52648c7b59615b2551de3eb1cd952118d7a27a24d5dede260b7f6703b"} Nov 25 14:00:49 crc kubenswrapper[4702]: I1125 14:00:49.585682 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sjhwl"] Nov 25 14:00:50 crc kubenswrapper[4702]: I1125 14:00:50.534554 4702 generic.go:334] "Generic (PLEG): container finished" podID="00f64e96-de2b-457c-a365-4dd090524959" containerID="9e057938b3682954f9b289acc75c8fdc4f4b25d765cdd37452f13bbbac9517dd" exitCode=0 Nov 25 14:00:50 crc kubenswrapper[4702]: I1125 14:00:50.534616 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r7t6l" event={"ID":"00f64e96-de2b-457c-a365-4dd090524959","Type":"ContainerDied","Data":"9e057938b3682954f9b289acc75c8fdc4f4b25d765cdd37452f13bbbac9517dd"} Nov 25 14:00:50 crc kubenswrapper[4702]: I1125 14:00:50.535452 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-sjhwl" podUID="147df53b-0c7b-4de1-b9b5-ee311584fdc7" containerName="registry-server" containerID="cri-o://69a14209d4fd25c41354e29adc203c641b675f84c612be6619177855e9dc0ea8" gracePeriod=2 Nov 25 14:00:52 crc kubenswrapper[4702]: I1125 14:00:52.681946 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"ac997ffc-32f2-4065-bc24-9982a8e306b6","Type":"ContainerStarted","Data":"ae2653573a1bd3c4c7a5631bbba1d16ae88fdac9b5a615f6a80d6969c7f320a4"} Nov 25 14:00:52 crc kubenswrapper[4702]: I1125 14:00:52.685051 4702 generic.go:334] "Generic (PLEG): container finished" podID="147df53b-0c7b-4de1-b9b5-ee311584fdc7" containerID="69a14209d4fd25c41354e29adc203c641b675f84c612be6619177855e9dc0ea8" exitCode=0 Nov 25 14:00:52 crc kubenswrapper[4702]: I1125 14:00:52.685118 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjhwl" event={"ID":"147df53b-0c7b-4de1-b9b5-ee311584fdc7","Type":"ContainerDied","Data":"69a14209d4fd25c41354e29adc203c641b675f84c612be6619177855e9dc0ea8"} Nov 25 14:00:52 crc kubenswrapper[4702]: I1125 14:00:52.979510 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sjhwl" Nov 25 14:00:53 crc kubenswrapper[4702]: I1125 14:00:53.073511 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/147df53b-0c7b-4de1-b9b5-ee311584fdc7-catalog-content\") pod \"147df53b-0c7b-4de1-b9b5-ee311584fdc7\" (UID: \"147df53b-0c7b-4de1-b9b5-ee311584fdc7\") " Nov 25 14:00:53 crc kubenswrapper[4702]: I1125 14:00:53.073609 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlvv9\" (UniqueName: \"kubernetes.io/projected/147df53b-0c7b-4de1-b9b5-ee311584fdc7-kube-api-access-vlvv9\") pod \"147df53b-0c7b-4de1-b9b5-ee311584fdc7\" (UID: \"147df53b-0c7b-4de1-b9b5-ee311584fdc7\") " Nov 25 14:00:53 crc kubenswrapper[4702]: I1125 14:00:53.073658 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/147df53b-0c7b-4de1-b9b5-ee311584fdc7-utilities\") pod \"147df53b-0c7b-4de1-b9b5-ee311584fdc7\" (UID: \"147df53b-0c7b-4de1-b9b5-ee311584fdc7\") " Nov 25 14:00:53 crc kubenswrapper[4702]: I1125 14:00:53.074665 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/147df53b-0c7b-4de1-b9b5-ee311584fdc7-utilities" (OuterVolumeSpecName: "utilities") pod "147df53b-0c7b-4de1-b9b5-ee311584fdc7" (UID: "147df53b-0c7b-4de1-b9b5-ee311584fdc7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:00:53 crc kubenswrapper[4702]: I1125 14:00:53.079913 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/147df53b-0c7b-4de1-b9b5-ee311584fdc7-kube-api-access-vlvv9" (OuterVolumeSpecName: "kube-api-access-vlvv9") pod "147df53b-0c7b-4de1-b9b5-ee311584fdc7" (UID: "147df53b-0c7b-4de1-b9b5-ee311584fdc7"). InnerVolumeSpecName "kube-api-access-vlvv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:00:53 crc kubenswrapper[4702]: I1125 14:00:53.123596 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/147df53b-0c7b-4de1-b9b5-ee311584fdc7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "147df53b-0c7b-4de1-b9b5-ee311584fdc7" (UID: "147df53b-0c7b-4de1-b9b5-ee311584fdc7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:00:53 crc kubenswrapper[4702]: I1125 14:00:53.174860 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/147df53b-0c7b-4de1-b9b5-ee311584fdc7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 14:00:53 crc kubenswrapper[4702]: I1125 14:00:53.174902 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlvv9\" (UniqueName: \"kubernetes.io/projected/147df53b-0c7b-4de1-b9b5-ee311584fdc7-kube-api-access-vlvv9\") on node \"crc\" DevicePath \"\"" Nov 25 14:00:53 crc kubenswrapper[4702]: I1125 14:00:53.174921 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/147df53b-0c7b-4de1-b9b5-ee311584fdc7-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 14:00:53 crc kubenswrapper[4702]: I1125 14:00:53.692654 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sjhwl" Nov 25 14:00:53 crc kubenswrapper[4702]: I1125 14:00:53.692667 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjhwl" event={"ID":"147df53b-0c7b-4de1-b9b5-ee311584fdc7","Type":"ContainerDied","Data":"de608e7d4dc0df04d87203bf82b5060d81878d3ab5fcc9ff5e3f979a01de8055"} Nov 25 14:00:53 crc kubenswrapper[4702]: I1125 14:00:53.692763 4702 scope.go:117] "RemoveContainer" containerID="69a14209d4fd25c41354e29adc203c641b675f84c612be6619177855e9dc0ea8" Nov 25 14:00:53 crc kubenswrapper[4702]: I1125 14:00:53.692907 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/memcached-0" Nov 25 14:00:53 crc kubenswrapper[4702]: I1125 14:00:53.714438 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/memcached-0" podStartSLOduration=3.149571915 podStartE2EDuration="14.714418336s" podCreationTimestamp="2025-11-25 14:00:39 +0000 UTC" firstStartedPulling="2025-11-25 14:00:40.615071711 +0000 UTC m=+849.777722171" lastFinishedPulling="2025-11-25 14:00:52.179918132 +0000 UTC m=+861.342568592" observedRunningTime="2025-11-25 14:00:53.710392266 +0000 UTC m=+862.873042726" watchObservedRunningTime="2025-11-25 14:00:53.714418336 +0000 UTC m=+862.877068786" Nov 25 14:00:53 crc kubenswrapper[4702]: I1125 14:00:53.731336 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sjhwl"] Nov 25 14:00:53 crc kubenswrapper[4702]: I1125 14:00:53.736379 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-sjhwl"] Nov 25 14:00:53 crc kubenswrapper[4702]: I1125 14:00:53.808070 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="147df53b-0c7b-4de1-b9b5-ee311584fdc7" path="/var/lib/kubelet/pods/147df53b-0c7b-4de1-b9b5-ee311584fdc7/volumes" Nov 25 14:00:53 crc kubenswrapper[4702]: I1125 14:00:53.833548 4702 scope.go:117] "RemoveContainer" containerID="75c4fd157d2d9ca8bf4819d111e99da9721f9123d040914affabb33c51bff4b7" Nov 25 14:00:53 crc kubenswrapper[4702]: I1125 14:00:53.980068 4702 scope.go:117] "RemoveContainer" containerID="068bd2be665b609399bac7c680646827506b632975549c74e0d791366d3402e1" Nov 25 14:00:54 crc kubenswrapper[4702]: I1125 14:00:54.203039 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-cf9vx"] Nov 25 14:00:54 crc kubenswrapper[4702]: E1125 14:00:54.203774 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="147df53b-0c7b-4de1-b9b5-ee311584fdc7" containerName="extract-content" Nov 25 14:00:54 crc kubenswrapper[4702]: I1125 14:00:54.203795 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="147df53b-0c7b-4de1-b9b5-ee311584fdc7" containerName="extract-content" Nov 25 14:00:54 crc kubenswrapper[4702]: E1125 14:00:54.203816 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="147df53b-0c7b-4de1-b9b5-ee311584fdc7" containerName="registry-server" Nov 25 14:00:54 crc kubenswrapper[4702]: I1125 14:00:54.203824 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="147df53b-0c7b-4de1-b9b5-ee311584fdc7" containerName="registry-server" Nov 25 14:00:54 crc kubenswrapper[4702]: E1125 14:00:54.203844 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="147df53b-0c7b-4de1-b9b5-ee311584fdc7" containerName="extract-utilities" Nov 25 14:00:54 crc kubenswrapper[4702]: I1125 14:00:54.203851 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="147df53b-0c7b-4de1-b9b5-ee311584fdc7" containerName="extract-utilities" Nov 25 14:00:54 crc kubenswrapper[4702]: I1125 14:00:54.203996 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="147df53b-0c7b-4de1-b9b5-ee311584fdc7" containerName="registry-server" Nov 25 14:00:54 crc kubenswrapper[4702]: I1125 14:00:54.207198 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cf9vx" Nov 25 14:00:54 crc kubenswrapper[4702]: I1125 14:00:54.223188 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cf9vx"] Nov 25 14:00:54 crc kubenswrapper[4702]: I1125 14:00:54.390724 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a7e12c8-26bc-4d1d-877c-88b89091b78e-catalog-content\") pod \"redhat-marketplace-cf9vx\" (UID: \"1a7e12c8-26bc-4d1d-877c-88b89091b78e\") " pod="openshift-marketplace/redhat-marketplace-cf9vx" Nov 25 14:00:54 crc kubenswrapper[4702]: I1125 14:00:54.390819 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwp2p\" (UniqueName: \"kubernetes.io/projected/1a7e12c8-26bc-4d1d-877c-88b89091b78e-kube-api-access-nwp2p\") pod \"redhat-marketplace-cf9vx\" (UID: \"1a7e12c8-26bc-4d1d-877c-88b89091b78e\") " pod="openshift-marketplace/redhat-marketplace-cf9vx" Nov 25 14:00:54 crc kubenswrapper[4702]: I1125 14:00:54.390875 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a7e12c8-26bc-4d1d-877c-88b89091b78e-utilities\") pod \"redhat-marketplace-cf9vx\" (UID: \"1a7e12c8-26bc-4d1d-877c-88b89091b78e\") " pod="openshift-marketplace/redhat-marketplace-cf9vx" Nov 25 14:00:54 crc kubenswrapper[4702]: I1125 14:00:54.492721 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a7e12c8-26bc-4d1d-877c-88b89091b78e-utilities\") pod \"redhat-marketplace-cf9vx\" (UID: \"1a7e12c8-26bc-4d1d-877c-88b89091b78e\") " pod="openshift-marketplace/redhat-marketplace-cf9vx" Nov 25 14:00:54 crc kubenswrapper[4702]: I1125 14:00:54.492781 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a7e12c8-26bc-4d1d-877c-88b89091b78e-catalog-content\") pod \"redhat-marketplace-cf9vx\" (UID: \"1a7e12c8-26bc-4d1d-877c-88b89091b78e\") " pod="openshift-marketplace/redhat-marketplace-cf9vx" Nov 25 14:00:54 crc kubenswrapper[4702]: I1125 14:00:54.492835 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwp2p\" (UniqueName: \"kubernetes.io/projected/1a7e12c8-26bc-4d1d-877c-88b89091b78e-kube-api-access-nwp2p\") pod \"redhat-marketplace-cf9vx\" (UID: \"1a7e12c8-26bc-4d1d-877c-88b89091b78e\") " pod="openshift-marketplace/redhat-marketplace-cf9vx" Nov 25 14:00:54 crc kubenswrapper[4702]: I1125 14:00:54.493395 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a7e12c8-26bc-4d1d-877c-88b89091b78e-utilities\") pod \"redhat-marketplace-cf9vx\" (UID: \"1a7e12c8-26bc-4d1d-877c-88b89091b78e\") " pod="openshift-marketplace/redhat-marketplace-cf9vx" Nov 25 14:00:54 crc kubenswrapper[4702]: I1125 14:00:54.493557 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a7e12c8-26bc-4d1d-877c-88b89091b78e-catalog-content\") pod \"redhat-marketplace-cf9vx\" (UID: \"1a7e12c8-26bc-4d1d-877c-88b89091b78e\") " pod="openshift-marketplace/redhat-marketplace-cf9vx" Nov 25 14:00:54 crc kubenswrapper[4702]: I1125 14:00:54.512338 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwp2p\" (UniqueName: \"kubernetes.io/projected/1a7e12c8-26bc-4d1d-877c-88b89091b78e-kube-api-access-nwp2p\") pod \"redhat-marketplace-cf9vx\" (UID: \"1a7e12c8-26bc-4d1d-877c-88b89091b78e\") " pod="openshift-marketplace/redhat-marketplace-cf9vx" Nov 25 14:00:54 crc kubenswrapper[4702]: I1125 14:00:54.539046 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cf9vx" Nov 25 14:00:54 crc kubenswrapper[4702]: I1125 14:00:54.708391 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r7t6l" event={"ID":"00f64e96-de2b-457c-a365-4dd090524959","Type":"ContainerStarted","Data":"23d178e2852da23b9f8bf326153032c80da016aaa4c63ab7ff58e8794b2cd3de"} Nov 25 14:00:54 crc kubenswrapper[4702]: I1125 14:00:54.852938 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cf9vx"] Nov 25 14:00:54 crc kubenswrapper[4702]: W1125 14:00:54.860963 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1a7e12c8_26bc_4d1d_877c_88b89091b78e.slice/crio-d2688086c0c8e87e96127304846ff5b5f621b61eb40ceb92f1a5bc8ada0a6840 WatchSource:0}: Error finding container d2688086c0c8e87e96127304846ff5b5f621b61eb40ceb92f1a5bc8ada0a6840: Status 404 returned error can't find the container with id d2688086c0c8e87e96127304846ff5b5f621b61eb40ceb92f1a5bc8ada0a6840 Nov 25 14:00:55 crc kubenswrapper[4702]: I1125 14:00:55.725652 4702 generic.go:334] "Generic (PLEG): container finished" podID="1a7e12c8-26bc-4d1d-877c-88b89091b78e" containerID="d4090e4bf5ce943ba63c157eaf77facb57583af02e12a89816809b15970d39a8" exitCode=0 Nov 25 14:00:55 crc kubenswrapper[4702]: I1125 14:00:55.725698 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cf9vx" event={"ID":"1a7e12c8-26bc-4d1d-877c-88b89091b78e","Type":"ContainerDied","Data":"d4090e4bf5ce943ba63c157eaf77facb57583af02e12a89816809b15970d39a8"} Nov 25 14:00:55 crc kubenswrapper[4702]: I1125 14:00:55.726047 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cf9vx" event={"ID":"1a7e12c8-26bc-4d1d-877c-88b89091b78e","Type":"ContainerStarted","Data":"d2688086c0c8e87e96127304846ff5b5f621b61eb40ceb92f1a5bc8ada0a6840"} Nov 25 14:00:55 crc kubenswrapper[4702]: I1125 14:00:55.728384 4702 generic.go:334] "Generic (PLEG): container finished" podID="00f64e96-de2b-457c-a365-4dd090524959" containerID="23d178e2852da23b9f8bf326153032c80da016aaa4c63ab7ff58e8794b2cd3de" exitCode=0 Nov 25 14:00:55 crc kubenswrapper[4702]: I1125 14:00:55.728458 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r7t6l" event={"ID":"00f64e96-de2b-457c-a365-4dd090524959","Type":"ContainerDied","Data":"23d178e2852da23b9f8bf326153032c80da016aaa4c63ab7ff58e8794b2cd3de"} Nov 25 14:00:58 crc kubenswrapper[4702]: I1125 14:00:58.755166 4702 generic.go:334] "Generic (PLEG): container finished" podID="f17baa10-42c4-4620-9b3c-a5dde9f78229" containerID="ac0330bdea5f28246e7216e4037f5e5ae1cb74c08cdc8de157d30c4676a16787" exitCode=0 Nov 25 14:00:58 crc kubenswrapper[4702]: I1125 14:00:58.755266 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"f17baa10-42c4-4620-9b3c-a5dde9f78229","Type":"ContainerDied","Data":"ac0330bdea5f28246e7216e4037f5e5ae1cb74c08cdc8de157d30c4676a16787"} Nov 25 14:00:58 crc kubenswrapper[4702]: I1125 14:00:58.760038 4702 generic.go:334] "Generic (PLEG): container finished" podID="bbd4f8e2-cb5f-45ff-ad8f-0faef399b018" containerID="4961265c62b5dc7b09f0fd529be6f7316c94e279538514c8bbffd00a07d6ab1d" exitCode=0 Nov 25 14:00:58 crc kubenswrapper[4702]: I1125 14:00:58.760070 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"bbd4f8e2-cb5f-45ff-ad8f-0faef399b018","Type":"ContainerDied","Data":"4961265c62b5dc7b09f0fd529be6f7316c94e279538514c8bbffd00a07d6ab1d"} Nov 25 14:00:59 crc kubenswrapper[4702]: I1125 14:00:59.768924 4702 generic.go:334] "Generic (PLEG): container finished" podID="1a7e12c8-26bc-4d1d-877c-88b89091b78e" containerID="ffddeb21b7ccf95c824e2e11378d2bfdea4f118d2fbca943d28503ef70669bf9" exitCode=0 Nov 25 14:00:59 crc kubenswrapper[4702]: I1125 14:00:59.769025 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cf9vx" event={"ID":"1a7e12c8-26bc-4d1d-877c-88b89091b78e","Type":"ContainerDied","Data":"ffddeb21b7ccf95c824e2e11378d2bfdea4f118d2fbca943d28503ef70669bf9"} Nov 25 14:00:59 crc kubenswrapper[4702]: I1125 14:00:59.773619 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r7t6l" event={"ID":"00f64e96-de2b-457c-a365-4dd090524959","Type":"ContainerStarted","Data":"faae125693f6ee75e4ddb3d7d76907288e4179f748ed8cda401eeceae0bfc160"} Nov 25 14:00:59 crc kubenswrapper[4702]: I1125 14:00:59.814304 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-r7t6l" podStartSLOduration=5.155371641 podStartE2EDuration="11.814282575s" podCreationTimestamp="2025-11-25 14:00:48 +0000 UTC" firstStartedPulling="2025-11-25 14:00:52.17222194 +0000 UTC m=+861.334872390" lastFinishedPulling="2025-11-25 14:00:58.831132874 +0000 UTC m=+867.993783324" observedRunningTime="2025-11-25 14:00:59.808289865 +0000 UTC m=+868.970940305" watchObservedRunningTime="2025-11-25 14:00:59.814282575 +0000 UTC m=+868.976933035" Nov 25 14:01:00 crc kubenswrapper[4702]: I1125 14:01:00.229423 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/memcached-0" Nov 25 14:01:00 crc kubenswrapper[4702]: I1125 14:01:00.810912 4702 generic.go:334] "Generic (PLEG): container finished" podID="8a586ccc-2c61-418c-a84c-51730c841205" containerID="608e3ff4bc93ddc6135275259aa92860868a9760b4ecfff2166ec4b862daafdf" exitCode=0 Nov 25 14:01:00 crc kubenswrapper[4702]: I1125 14:01:00.811042 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"8a586ccc-2c61-418c-a84c-51730c841205","Type":"ContainerDied","Data":"608e3ff4bc93ddc6135275259aa92860868a9760b4ecfff2166ec4b862daafdf"} Nov 25 14:01:01 crc kubenswrapper[4702]: I1125 14:01:01.820989 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-9wl6w" event={"ID":"fc79fae5-b7b4-43b6-89ad-90008bf8dd57","Type":"ContainerStarted","Data":"69b0c944e92c8d4fe6ba6a841269049751c6952aeb80f74711ddfa9a3b9c9687"} Nov 25 14:01:01 crc kubenswrapper[4702]: I1125 14:01:01.823932 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"f17baa10-42c4-4620-9b3c-a5dde9f78229","Type":"ContainerStarted","Data":"3b10bbb0afa913749d594b3376f53b5b91ee395ad8742b9f44991ac252b453b7"} Nov 25 14:01:01 crc kubenswrapper[4702]: I1125 14:01:01.826832 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"bbd4f8e2-cb5f-45ff-ad8f-0faef399b018","Type":"ContainerStarted","Data":"aa220b295a4bb36846ea6fecf2948894d46de809b17008b0637d67bf26e1d8e0"} Nov 25 14:01:01 crc kubenswrapper[4702]: I1125 14:01:01.829544 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"8a586ccc-2c61-418c-a84c-51730c841205","Type":"ContainerStarted","Data":"d6a8d8937cdff32278eaded2a09b5b10e44291365a217e7fd230542dd8237e9f"} Nov 25 14:01:01 crc kubenswrapper[4702]: I1125 14:01:01.832094 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cf9vx" event={"ID":"1a7e12c8-26bc-4d1d-877c-88b89091b78e","Type":"ContainerStarted","Data":"7a0cb0aac2f05c0b94b27450cab3f1b77c6ca3823e0ccb9f0996de63481ed671"} Nov 25 14:01:01 crc kubenswrapper[4702]: I1125 14:01:01.843913 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-9wl6w" podStartSLOduration=3.942193102 podStartE2EDuration="18.843891232s" podCreationTimestamp="2025-11-25 14:00:43 +0000 UTC" firstStartedPulling="2025-11-25 14:00:45.859072767 +0000 UTC m=+855.021723217" lastFinishedPulling="2025-11-25 14:01:00.760770897 +0000 UTC m=+869.923421347" observedRunningTime="2025-11-25 14:01:01.839831051 +0000 UTC m=+871.002481501" watchObservedRunningTime="2025-11-25 14:01:01.843891232 +0000 UTC m=+871.006541682" Nov 25 14:01:01 crc kubenswrapper[4702]: I1125 14:01:01.865895 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-cf9vx" podStartSLOduration=2.695005058 podStartE2EDuration="7.865867812s" podCreationTimestamp="2025-11-25 14:00:54 +0000 UTC" firstStartedPulling="2025-11-25 14:00:55.772346473 +0000 UTC m=+864.934996963" lastFinishedPulling="2025-11-25 14:01:00.943209277 +0000 UTC m=+870.105859717" observedRunningTime="2025-11-25 14:01:01.861568395 +0000 UTC m=+871.024218855" watchObservedRunningTime="2025-11-25 14:01:01.865867812 +0000 UTC m=+871.028518262" Nov 25 14:01:01 crc kubenswrapper[4702]: I1125 14:01:01.885612 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-0" podStartSLOduration=25.676761613 podStartE2EDuration="1m1.885592646s" podCreationTimestamp="2025-11-25 14:00:00 +0000 UTC" firstStartedPulling="2025-11-25 14:00:03.090117115 +0000 UTC m=+812.252767565" lastFinishedPulling="2025-11-25 14:00:39.298948148 +0000 UTC m=+848.461598598" observedRunningTime="2025-11-25 14:01:01.881293279 +0000 UTC m=+871.043943729" watchObservedRunningTime="2025-11-25 14:01:01.885592646 +0000 UTC m=+871.048243096" Nov 25 14:01:01 crc kubenswrapper[4702]: I1125 14:01:01.908034 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-1" podStartSLOduration=-9223371974.946768 podStartE2EDuration="1m1.908008828s" podCreationTimestamp="2025-11-25 14:00:00 +0000 UTC" firstStartedPulling="2025-11-25 14:00:02.926116436 +0000 UTC m=+812.088766886" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 14:01:01.903206928 +0000 UTC m=+871.065857398" watchObservedRunningTime="2025-11-25 14:01:01.908008828 +0000 UTC m=+871.070659278" Nov 25 14:01:01 crc kubenswrapper[4702]: I1125 14:01:01.924452 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-2" podStartSLOduration=25.438487811999998 podStartE2EDuration="1m1.92442965s" podCreationTimestamp="2025-11-25 14:00:00 +0000 UTC" firstStartedPulling="2025-11-25 14:00:02.811965874 +0000 UTC m=+811.974616324" lastFinishedPulling="2025-11-25 14:00:39.297907692 +0000 UTC m=+848.460558162" observedRunningTime="2025-11-25 14:01:01.923597689 +0000 UTC m=+871.086248139" watchObservedRunningTime="2025-11-25 14:01:01.92442965 +0000 UTC m=+871.087080100" Nov 25 14:01:02 crc kubenswrapper[4702]: I1125 14:01:02.305406 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 14:01:02 crc kubenswrapper[4702]: I1125 14:01:02.305478 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 14:01:02 crc kubenswrapper[4702]: I1125 14:01:02.322106 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 14:01:02 crc kubenswrapper[4702]: I1125 14:01:02.322145 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 14:01:02 crc kubenswrapper[4702]: I1125 14:01:02.344915 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 14:01:02 crc kubenswrapper[4702]: I1125 14:01:02.344973 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 14:01:04 crc kubenswrapper[4702]: I1125 14:01:04.108083 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/rabbitmq-cluster-operator-index-9wl6w" Nov 25 14:01:04 crc kubenswrapper[4702]: I1125 14:01:04.108552 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/rabbitmq-cluster-operator-index-9wl6w" Nov 25 14:01:04 crc kubenswrapper[4702]: I1125 14:01:04.151052 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/rabbitmq-cluster-operator-index-9wl6w" Nov 25 14:01:04 crc kubenswrapper[4702]: I1125 14:01:04.539887 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-cf9vx" Nov 25 14:01:04 crc kubenswrapper[4702]: I1125 14:01:04.540319 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-cf9vx" Nov 25 14:01:04 crc kubenswrapper[4702]: I1125 14:01:04.584090 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-cf9vx" Nov 25 14:01:04 crc kubenswrapper[4702]: I1125 14:01:04.797048 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vkrfk"] Nov 25 14:01:04 crc kubenswrapper[4702]: I1125 14:01:04.799451 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vkrfk" Nov 25 14:01:04 crc kubenswrapper[4702]: I1125 14:01:04.818024 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vkrfk"] Nov 25 14:01:04 crc kubenswrapper[4702]: I1125 14:01:04.858669 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a-catalog-content\") pod \"redhat-operators-vkrfk\" (UID: \"815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a\") " pod="openshift-marketplace/redhat-operators-vkrfk" Nov 25 14:01:04 crc kubenswrapper[4702]: I1125 14:01:04.858780 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a-utilities\") pod \"redhat-operators-vkrfk\" (UID: \"815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a\") " pod="openshift-marketplace/redhat-operators-vkrfk" Nov 25 14:01:04 crc kubenswrapper[4702]: I1125 14:01:04.858819 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sj8db\" (UniqueName: \"kubernetes.io/projected/815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a-kube-api-access-sj8db\") pod \"redhat-operators-vkrfk\" (UID: \"815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a\") " pod="openshift-marketplace/redhat-operators-vkrfk" Nov 25 14:01:04 crc kubenswrapper[4702]: I1125 14:01:04.960553 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a-utilities\") pod \"redhat-operators-vkrfk\" (UID: \"815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a\") " pod="openshift-marketplace/redhat-operators-vkrfk" Nov 25 14:01:04 crc kubenswrapper[4702]: I1125 14:01:04.960614 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sj8db\" (UniqueName: \"kubernetes.io/projected/815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a-kube-api-access-sj8db\") pod \"redhat-operators-vkrfk\" (UID: \"815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a\") " pod="openshift-marketplace/redhat-operators-vkrfk" Nov 25 14:01:04 crc kubenswrapper[4702]: I1125 14:01:04.960701 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a-catalog-content\") pod \"redhat-operators-vkrfk\" (UID: \"815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a\") " pod="openshift-marketplace/redhat-operators-vkrfk" Nov 25 14:01:04 crc kubenswrapper[4702]: I1125 14:01:04.961221 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a-utilities\") pod \"redhat-operators-vkrfk\" (UID: \"815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a\") " pod="openshift-marketplace/redhat-operators-vkrfk" Nov 25 14:01:04 crc kubenswrapper[4702]: I1125 14:01:04.961273 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a-catalog-content\") pod \"redhat-operators-vkrfk\" (UID: \"815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a\") " pod="openshift-marketplace/redhat-operators-vkrfk" Nov 25 14:01:04 crc kubenswrapper[4702]: I1125 14:01:04.983103 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sj8db\" (UniqueName: \"kubernetes.io/projected/815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a-kube-api-access-sj8db\") pod \"redhat-operators-vkrfk\" (UID: \"815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a\") " pod="openshift-marketplace/redhat-operators-vkrfk" Nov 25 14:01:05 crc kubenswrapper[4702]: I1125 14:01:05.129805 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vkrfk" Nov 25 14:01:05 crc kubenswrapper[4702]: I1125 14:01:05.610936 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vkrfk"] Nov 25 14:01:05 crc kubenswrapper[4702]: I1125 14:01:05.871115 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vkrfk" event={"ID":"815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a","Type":"ContainerStarted","Data":"56986cd6ae59e6ad4b31aee320060f3fc527aa010d9887534d223f9ff4916626"} Nov 25 14:01:06 crc kubenswrapper[4702]: I1125 14:01:06.878040 4702 generic.go:334] "Generic (PLEG): container finished" podID="815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a" containerID="2b311669ef89fab371689904adeadeb2d5759ea31eec30396079cad98594e141" exitCode=0 Nov 25 14:01:06 crc kubenswrapper[4702]: I1125 14:01:06.878093 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vkrfk" event={"ID":"815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a","Type":"ContainerDied","Data":"2b311669ef89fab371689904adeadeb2d5759ea31eec30396079cad98594e141"} Nov 25 14:01:08 crc kubenswrapper[4702]: I1125 14:01:08.900828 4702 generic.go:334] "Generic (PLEG): container finished" podID="815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a" containerID="102b681c7b99440e1cc3b3aff724d38604f2c3ccd2babbdd6c9531039183cdcd" exitCode=0 Nov 25 14:01:08 crc kubenswrapper[4702]: I1125 14:01:08.900904 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vkrfk" event={"ID":"815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a","Type":"ContainerDied","Data":"102b681c7b99440e1cc3b3aff724d38604f2c3ccd2babbdd6c9531039183cdcd"} Nov 25 14:01:09 crc kubenswrapper[4702]: I1125 14:01:09.122632 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-r7t6l" Nov 25 14:01:09 crc kubenswrapper[4702]: I1125 14:01:09.122697 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-r7t6l" Nov 25 14:01:09 crc kubenswrapper[4702]: I1125 14:01:09.168409 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-r7t6l" Nov 25 14:01:09 crc kubenswrapper[4702]: E1125 14:01:09.406100 4702 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.180:50798->38.102.83.180:45641: write tcp 38.102.83.180:50798->38.102.83.180:45641: write: broken pipe Nov 25 14:01:09 crc kubenswrapper[4702]: I1125 14:01:09.911886 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vkrfk" event={"ID":"815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a","Type":"ContainerStarted","Data":"bd8c1c946f52c57e4dd165778dbaf1a597d5d38333b000ea7c48c5588623297e"} Nov 25 14:01:09 crc kubenswrapper[4702]: I1125 14:01:09.936151 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vkrfk" podStartSLOduration=3.477785126 podStartE2EDuration="5.936121645s" podCreationTimestamp="2025-11-25 14:01:04 +0000 UTC" firstStartedPulling="2025-11-25 14:01:06.880066212 +0000 UTC m=+876.042716662" lastFinishedPulling="2025-11-25 14:01:09.338402721 +0000 UTC m=+878.501053181" observedRunningTime="2025-11-25 14:01:09.93232568 +0000 UTC m=+879.094976150" watchObservedRunningTime="2025-11-25 14:01:09.936121645 +0000 UTC m=+879.098772095" Nov 25 14:01:09 crc kubenswrapper[4702]: I1125 14:01:09.965028 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-r7t6l" Nov 25 14:01:12 crc kubenswrapper[4702]: I1125 14:01:12.447196 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 14:01:12 crc kubenswrapper[4702]: I1125 14:01:12.538103 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 14:01:12 crc kubenswrapper[4702]: I1125 14:01:12.595064 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r7t6l"] Nov 25 14:01:12 crc kubenswrapper[4702]: I1125 14:01:12.595429 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-r7t6l" podUID="00f64e96-de2b-457c-a365-4dd090524959" containerName="registry-server" containerID="cri-o://faae125693f6ee75e4ddb3d7d76907288e4179f748ed8cda401eeceae0bfc160" gracePeriod=2 Nov 25 14:01:12 crc kubenswrapper[4702]: E1125 14:01:12.783336 4702 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00f64e96_de2b_457c_a365_4dd090524959.slice/crio-faae125693f6ee75e4ddb3d7d76907288e4179f748ed8cda401eeceae0bfc160.scope\": RecentStats: unable to find data in memory cache]" Nov 25 14:01:13 crc kubenswrapper[4702]: I1125 14:01:12.937019 4702 generic.go:334] "Generic (PLEG): container finished" podID="00f64e96-de2b-457c-a365-4dd090524959" containerID="faae125693f6ee75e4ddb3d7d76907288e4179f748ed8cda401eeceae0bfc160" exitCode=0 Nov 25 14:01:13 crc kubenswrapper[4702]: I1125 14:01:12.938098 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r7t6l" event={"ID":"00f64e96-de2b-457c-a365-4dd090524959","Type":"ContainerDied","Data":"faae125693f6ee75e4ddb3d7d76907288e4179f748ed8cda401eeceae0bfc160"} Nov 25 14:01:13 crc kubenswrapper[4702]: E1125 14:01:13.015196 4702 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.180:45020->38.102.83.180:45641: write tcp 38.102.83.180:45020->38.102.83.180:45641: write: broken pipe Nov 25 14:01:13 crc kubenswrapper[4702]: I1125 14:01:13.065601 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r7t6l" Nov 25 14:01:13 crc kubenswrapper[4702]: I1125 14:01:13.217060 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00f64e96-de2b-457c-a365-4dd090524959-utilities\") pod \"00f64e96-de2b-457c-a365-4dd090524959\" (UID: \"00f64e96-de2b-457c-a365-4dd090524959\") " Nov 25 14:01:13 crc kubenswrapper[4702]: I1125 14:01:13.217208 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00f64e96-de2b-457c-a365-4dd090524959-catalog-content\") pod \"00f64e96-de2b-457c-a365-4dd090524959\" (UID: \"00f64e96-de2b-457c-a365-4dd090524959\") " Nov 25 14:01:13 crc kubenswrapper[4702]: I1125 14:01:13.217343 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbl8t\" (UniqueName: \"kubernetes.io/projected/00f64e96-de2b-457c-a365-4dd090524959-kube-api-access-cbl8t\") pod \"00f64e96-de2b-457c-a365-4dd090524959\" (UID: \"00f64e96-de2b-457c-a365-4dd090524959\") " Nov 25 14:01:13 crc kubenswrapper[4702]: I1125 14:01:13.218592 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00f64e96-de2b-457c-a365-4dd090524959-utilities" (OuterVolumeSpecName: "utilities") pod "00f64e96-de2b-457c-a365-4dd090524959" (UID: "00f64e96-de2b-457c-a365-4dd090524959"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:01:13 crc kubenswrapper[4702]: I1125 14:01:13.226700 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00f64e96-de2b-457c-a365-4dd090524959-kube-api-access-cbl8t" (OuterVolumeSpecName: "kube-api-access-cbl8t") pod "00f64e96-de2b-457c-a365-4dd090524959" (UID: "00f64e96-de2b-457c-a365-4dd090524959"). InnerVolumeSpecName "kube-api-access-cbl8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:01:13 crc kubenswrapper[4702]: I1125 14:01:13.282101 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00f64e96-de2b-457c-a365-4dd090524959-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "00f64e96-de2b-457c-a365-4dd090524959" (UID: "00f64e96-de2b-457c-a365-4dd090524959"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:01:13 crc kubenswrapper[4702]: I1125 14:01:13.319512 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00f64e96-de2b-457c-a365-4dd090524959-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 14:01:13 crc kubenswrapper[4702]: I1125 14:01:13.319552 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00f64e96-de2b-457c-a365-4dd090524959-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 14:01:13 crc kubenswrapper[4702]: I1125 14:01:13.319566 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbl8t\" (UniqueName: \"kubernetes.io/projected/00f64e96-de2b-457c-a365-4dd090524959-kube-api-access-cbl8t\") on node \"crc\" DevicePath \"\"" Nov 25 14:01:13 crc kubenswrapper[4702]: I1125 14:01:13.946872 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r7t6l" event={"ID":"00f64e96-de2b-457c-a365-4dd090524959","Type":"ContainerDied","Data":"534685b52648c7b59615b2551de3eb1cd952118d7a27a24d5dede260b7f6703b"} Nov 25 14:01:13 crc kubenswrapper[4702]: I1125 14:01:13.946951 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r7t6l" Nov 25 14:01:13 crc kubenswrapper[4702]: I1125 14:01:13.946959 4702 scope.go:117] "RemoveContainer" containerID="faae125693f6ee75e4ddb3d7d76907288e4179f748ed8cda401eeceae0bfc160" Nov 25 14:01:13 crc kubenswrapper[4702]: I1125 14:01:13.966274 4702 scope.go:117] "RemoveContainer" containerID="23d178e2852da23b9f8bf326153032c80da016aaa4c63ab7ff58e8794b2cd3de" Nov 25 14:01:13 crc kubenswrapper[4702]: I1125 14:01:13.969888 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r7t6l"] Nov 25 14:01:13 crc kubenswrapper[4702]: I1125 14:01:13.981185 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-r7t6l"] Nov 25 14:01:13 crc kubenswrapper[4702]: I1125 14:01:13.989782 4702 scope.go:117] "RemoveContainer" containerID="9e057938b3682954f9b289acc75c8fdc4f4b25d765cdd37452f13bbbac9517dd" Nov 25 14:01:14 crc kubenswrapper[4702]: I1125 14:01:14.160253 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/rabbitmq-cluster-operator-index-9wl6w" Nov 25 14:01:14 crc kubenswrapper[4702]: I1125 14:01:14.586153 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-cf9vx" Nov 25 14:01:15 crc kubenswrapper[4702]: I1125 14:01:15.130065 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vkrfk" Nov 25 14:01:15 crc kubenswrapper[4702]: I1125 14:01:15.130160 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vkrfk" Nov 25 14:01:15 crc kubenswrapper[4702]: I1125 14:01:15.181550 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vkrfk" Nov 25 14:01:15 crc kubenswrapper[4702]: I1125 14:01:15.813550 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00f64e96-de2b-457c-a365-4dd090524959" path="/var/lib/kubelet/pods/00f64e96-de2b-457c-a365-4dd090524959/volumes" Nov 25 14:01:16 crc kubenswrapper[4702]: I1125 14:01:16.057679 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vkrfk" Nov 25 14:01:17 crc kubenswrapper[4702]: I1125 14:01:17.053996 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5"] Nov 25 14:01:17 crc kubenswrapper[4702]: E1125 14:01:17.054356 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00f64e96-de2b-457c-a365-4dd090524959" containerName="registry-server" Nov 25 14:01:17 crc kubenswrapper[4702]: I1125 14:01:17.054374 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="00f64e96-de2b-457c-a365-4dd090524959" containerName="registry-server" Nov 25 14:01:17 crc kubenswrapper[4702]: E1125 14:01:17.054391 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00f64e96-de2b-457c-a365-4dd090524959" containerName="extract-content" Nov 25 14:01:17 crc kubenswrapper[4702]: I1125 14:01:17.054400 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="00f64e96-de2b-457c-a365-4dd090524959" containerName="extract-content" Nov 25 14:01:17 crc kubenswrapper[4702]: E1125 14:01:17.054415 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00f64e96-de2b-457c-a365-4dd090524959" containerName="extract-utilities" Nov 25 14:01:17 crc kubenswrapper[4702]: I1125 14:01:17.054423 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="00f64e96-de2b-457c-a365-4dd090524959" containerName="extract-utilities" Nov 25 14:01:17 crc kubenswrapper[4702]: I1125 14:01:17.054574 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="00f64e96-de2b-457c-a365-4dd090524959" containerName="registry-server" Nov 25 14:01:17 crc kubenswrapper[4702]: I1125 14:01:17.055761 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5" Nov 25 14:01:17 crc kubenswrapper[4702]: I1125 14:01:17.058589 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-6r4jq" Nov 25 14:01:17 crc kubenswrapper[4702]: I1125 14:01:17.072151 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5"] Nov 25 14:01:17 crc kubenswrapper[4702]: I1125 14:01:17.080999 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/aeb59e3f-26d5-426a-a371-cdc4efd75add-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5\" (UID: \"aeb59e3f-26d5-426a-a371-cdc4efd75add\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5" Nov 25 14:01:17 crc kubenswrapper[4702]: I1125 14:01:17.081155 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjm8x\" (UniqueName: \"kubernetes.io/projected/aeb59e3f-26d5-426a-a371-cdc4efd75add-kube-api-access-hjm8x\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5\" (UID: \"aeb59e3f-26d5-426a-a371-cdc4efd75add\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5" Nov 25 14:01:17 crc kubenswrapper[4702]: I1125 14:01:17.081181 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/aeb59e3f-26d5-426a-a371-cdc4efd75add-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5\" (UID: \"aeb59e3f-26d5-426a-a371-cdc4efd75add\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5" Nov 25 14:01:17 crc kubenswrapper[4702]: I1125 14:01:17.182787 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjm8x\" (UniqueName: \"kubernetes.io/projected/aeb59e3f-26d5-426a-a371-cdc4efd75add-kube-api-access-hjm8x\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5\" (UID: \"aeb59e3f-26d5-426a-a371-cdc4efd75add\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5" Nov 25 14:01:17 crc kubenswrapper[4702]: I1125 14:01:17.182871 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/aeb59e3f-26d5-426a-a371-cdc4efd75add-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5\" (UID: \"aeb59e3f-26d5-426a-a371-cdc4efd75add\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5" Nov 25 14:01:17 crc kubenswrapper[4702]: I1125 14:01:17.182937 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/aeb59e3f-26d5-426a-a371-cdc4efd75add-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5\" (UID: \"aeb59e3f-26d5-426a-a371-cdc4efd75add\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5" Nov 25 14:01:17 crc kubenswrapper[4702]: I1125 14:01:17.183399 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/aeb59e3f-26d5-426a-a371-cdc4efd75add-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5\" (UID: \"aeb59e3f-26d5-426a-a371-cdc4efd75add\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5" Nov 25 14:01:17 crc kubenswrapper[4702]: I1125 14:01:17.183494 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/aeb59e3f-26d5-426a-a371-cdc4efd75add-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5\" (UID: \"aeb59e3f-26d5-426a-a371-cdc4efd75add\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5" Nov 25 14:01:17 crc kubenswrapper[4702]: I1125 14:01:17.204744 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjm8x\" (UniqueName: \"kubernetes.io/projected/aeb59e3f-26d5-426a-a371-cdc4efd75add-kube-api-access-hjm8x\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5\" (UID: \"aeb59e3f-26d5-426a-a371-cdc4efd75add\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5" Nov 25 14:01:17 crc kubenswrapper[4702]: I1125 14:01:17.422356 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5" Nov 25 14:01:17 crc kubenswrapper[4702]: I1125 14:01:17.863127 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5"] Nov 25 14:01:17 crc kubenswrapper[4702]: I1125 14:01:17.979669 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5" event={"ID":"aeb59e3f-26d5-426a-a371-cdc4efd75add","Type":"ContainerStarted","Data":"0c323de7ce63c0fbdc7978a4bb8416a3d684ce193839ba035d396d2a4ebd688f"} Nov 25 14:01:18 crc kubenswrapper[4702]: I1125 14:01:18.989888 4702 generic.go:334] "Generic (PLEG): container finished" podID="aeb59e3f-26d5-426a-a371-cdc4efd75add" containerID="b837ef0e1bedb3c37ee85fe9acb95b0ccf3c6be88d6e919e346f3a81c73835da" exitCode=0 Nov 25 14:01:18 crc kubenswrapper[4702]: I1125 14:01:18.989937 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5" event={"ID":"aeb59e3f-26d5-426a-a371-cdc4efd75add","Type":"ContainerDied","Data":"b837ef0e1bedb3c37ee85fe9acb95b0ccf3c6be88d6e919e346f3a81c73835da"} Nov 25 14:01:22 crc kubenswrapper[4702]: I1125 14:01:22.015810 4702 generic.go:334] "Generic (PLEG): container finished" podID="aeb59e3f-26d5-426a-a371-cdc4efd75add" containerID="83f172510f3abd5f67d67e125303429e121bb24b260e4941277553d252753983" exitCode=0 Nov 25 14:01:22 crc kubenswrapper[4702]: I1125 14:01:22.015945 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5" event={"ID":"aeb59e3f-26d5-426a-a371-cdc4efd75add","Type":"ContainerDied","Data":"83f172510f3abd5f67d67e125303429e121bb24b260e4941277553d252753983"} Nov 25 14:01:22 crc kubenswrapper[4702]: I1125 14:01:22.384690 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/openstack-galera-2" podUID="bbd4f8e2-cb5f-45ff-ad8f-0faef399b018" containerName="galera" probeResult="failure" output=< Nov 25 14:01:22 crc kubenswrapper[4702]: wsrep_local_state_comment (Donor/Desynced) differs from Synced Nov 25 14:01:22 crc kubenswrapper[4702]: > Nov 25 14:01:23 crc kubenswrapper[4702]: I1125 14:01:23.027647 4702 generic.go:334] "Generic (PLEG): container finished" podID="aeb59e3f-26d5-426a-a371-cdc4efd75add" containerID="9ccaf74d61a1886f58e23aae895f002940b362f5dc6b9f451d7b749cbfc667fe" exitCode=0 Nov 25 14:01:23 crc kubenswrapper[4702]: I1125 14:01:23.027754 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5" event={"ID":"aeb59e3f-26d5-426a-a371-cdc4efd75add","Type":"ContainerDied","Data":"9ccaf74d61a1886f58e23aae895f002940b362f5dc6b9f451d7b749cbfc667fe"} Nov 25 14:01:23 crc kubenswrapper[4702]: I1125 14:01:23.184970 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cf9vx"] Nov 25 14:01:23 crc kubenswrapper[4702]: I1125 14:01:23.185369 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-cf9vx" podUID="1a7e12c8-26bc-4d1d-877c-88b89091b78e" containerName="registry-server" containerID="cri-o://7a0cb0aac2f05c0b94b27450cab3f1b77c6ca3823e0ccb9f0996de63481ed671" gracePeriod=2 Nov 25 14:01:23 crc kubenswrapper[4702]: I1125 14:01:23.589426 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vkrfk"] Nov 25 14:01:23 crc kubenswrapper[4702]: I1125 14:01:23.590318 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vkrfk" podUID="815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a" containerName="registry-server" containerID="cri-o://bd8c1c946f52c57e4dd165778dbaf1a597d5d38333b000ea7c48c5588623297e" gracePeriod=2 Nov 25 14:01:23 crc kubenswrapper[4702]: I1125 14:01:23.794951 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cf9vx" Nov 25 14:01:23 crc kubenswrapper[4702]: I1125 14:01:23.891471 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a7e12c8-26bc-4d1d-877c-88b89091b78e-utilities\") pod \"1a7e12c8-26bc-4d1d-877c-88b89091b78e\" (UID: \"1a7e12c8-26bc-4d1d-877c-88b89091b78e\") " Nov 25 14:01:23 crc kubenswrapper[4702]: I1125 14:01:23.891585 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nwp2p\" (UniqueName: \"kubernetes.io/projected/1a7e12c8-26bc-4d1d-877c-88b89091b78e-kube-api-access-nwp2p\") pod \"1a7e12c8-26bc-4d1d-877c-88b89091b78e\" (UID: \"1a7e12c8-26bc-4d1d-877c-88b89091b78e\") " Nov 25 14:01:23 crc kubenswrapper[4702]: I1125 14:01:23.891674 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a7e12c8-26bc-4d1d-877c-88b89091b78e-catalog-content\") pod \"1a7e12c8-26bc-4d1d-877c-88b89091b78e\" (UID: \"1a7e12c8-26bc-4d1d-877c-88b89091b78e\") " Nov 25 14:01:23 crc kubenswrapper[4702]: I1125 14:01:23.892816 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a7e12c8-26bc-4d1d-877c-88b89091b78e-utilities" (OuterVolumeSpecName: "utilities") pod "1a7e12c8-26bc-4d1d-877c-88b89091b78e" (UID: "1a7e12c8-26bc-4d1d-877c-88b89091b78e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:01:23 crc kubenswrapper[4702]: I1125 14:01:23.914797 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a7e12c8-26bc-4d1d-877c-88b89091b78e-kube-api-access-nwp2p" (OuterVolumeSpecName: "kube-api-access-nwp2p") pod "1a7e12c8-26bc-4d1d-877c-88b89091b78e" (UID: "1a7e12c8-26bc-4d1d-877c-88b89091b78e"). InnerVolumeSpecName "kube-api-access-nwp2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:01:23 crc kubenswrapper[4702]: I1125 14:01:23.918003 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a7e12c8-26bc-4d1d-877c-88b89091b78e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1a7e12c8-26bc-4d1d-877c-88b89091b78e" (UID: "1a7e12c8-26bc-4d1d-877c-88b89091b78e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:01:23 crc kubenswrapper[4702]: I1125 14:01:23.993442 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a7e12c8-26bc-4d1d-877c-88b89091b78e-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 14:01:23 crc kubenswrapper[4702]: I1125 14:01:23.993496 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nwp2p\" (UniqueName: \"kubernetes.io/projected/1a7e12c8-26bc-4d1d-877c-88b89091b78e-kube-api-access-nwp2p\") on node \"crc\" DevicePath \"\"" Nov 25 14:01:23 crc kubenswrapper[4702]: I1125 14:01:23.993514 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a7e12c8-26bc-4d1d-877c-88b89091b78e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.025558 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vkrfk" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.037217 4702 generic.go:334] "Generic (PLEG): container finished" podID="815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a" containerID="bd8c1c946f52c57e4dd165778dbaf1a597d5d38333b000ea7c48c5588623297e" exitCode=0 Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.037308 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vkrfk" event={"ID":"815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a","Type":"ContainerDied","Data":"bd8c1c946f52c57e4dd165778dbaf1a597d5d38333b000ea7c48c5588623297e"} Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.037346 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vkrfk" event={"ID":"815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a","Type":"ContainerDied","Data":"56986cd6ae59e6ad4b31aee320060f3fc527aa010d9887534d223f9ff4916626"} Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.037369 4702 scope.go:117] "RemoveContainer" containerID="bd8c1c946f52c57e4dd165778dbaf1a597d5d38333b000ea7c48c5588623297e" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.037552 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vkrfk" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.042807 4702 generic.go:334] "Generic (PLEG): container finished" podID="1a7e12c8-26bc-4d1d-877c-88b89091b78e" containerID="7a0cb0aac2f05c0b94b27450cab3f1b77c6ca3823e0ccb9f0996de63481ed671" exitCode=0 Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.042876 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cf9vx" event={"ID":"1a7e12c8-26bc-4d1d-877c-88b89091b78e","Type":"ContainerDied","Data":"7a0cb0aac2f05c0b94b27450cab3f1b77c6ca3823e0ccb9f0996de63481ed671"} Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.042922 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cf9vx" event={"ID":"1a7e12c8-26bc-4d1d-877c-88b89091b78e","Type":"ContainerDied","Data":"d2688086c0c8e87e96127304846ff5b5f621b61eb40ceb92f1a5bc8ada0a6840"} Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.042933 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cf9vx" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.095005 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sj8db\" (UniqueName: \"kubernetes.io/projected/815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a-kube-api-access-sj8db\") pod \"815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a\" (UID: \"815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a\") " Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.095094 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a-catalog-content\") pod \"815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a\" (UID: \"815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a\") " Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.095160 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a-utilities\") pod \"815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a\" (UID: \"815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a\") " Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.103158 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a-utilities" (OuterVolumeSpecName: "utilities") pod "815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a" (UID: "815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.105086 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cf9vx"] Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.106181 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a-kube-api-access-sj8db" (OuterVolumeSpecName: "kube-api-access-sj8db") pod "815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a" (UID: "815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a"). InnerVolumeSpecName "kube-api-access-sj8db". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.111471 4702 scope.go:117] "RemoveContainer" containerID="102b681c7b99440e1cc3b3aff724d38604f2c3ccd2babbdd6c9531039183cdcd" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.111963 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-cf9vx"] Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.142548 4702 scope.go:117] "RemoveContainer" containerID="2b311669ef89fab371689904adeadeb2d5759ea31eec30396079cad98594e141" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.183391 4702 scope.go:117] "RemoveContainer" containerID="bd8c1c946f52c57e4dd165778dbaf1a597d5d38333b000ea7c48c5588623297e" Nov 25 14:01:24 crc kubenswrapper[4702]: E1125 14:01:24.189931 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd8c1c946f52c57e4dd165778dbaf1a597d5d38333b000ea7c48c5588623297e\": container with ID starting with bd8c1c946f52c57e4dd165778dbaf1a597d5d38333b000ea7c48c5588623297e not found: ID does not exist" containerID="bd8c1c946f52c57e4dd165778dbaf1a597d5d38333b000ea7c48c5588623297e" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.189991 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd8c1c946f52c57e4dd165778dbaf1a597d5d38333b000ea7c48c5588623297e"} err="failed to get container status \"bd8c1c946f52c57e4dd165778dbaf1a597d5d38333b000ea7c48c5588623297e\": rpc error: code = NotFound desc = could not find container \"bd8c1c946f52c57e4dd165778dbaf1a597d5d38333b000ea7c48c5588623297e\": container with ID starting with bd8c1c946f52c57e4dd165778dbaf1a597d5d38333b000ea7c48c5588623297e not found: ID does not exist" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.190029 4702 scope.go:117] "RemoveContainer" containerID="102b681c7b99440e1cc3b3aff724d38604f2c3ccd2babbdd6c9531039183cdcd" Nov 25 14:01:24 crc kubenswrapper[4702]: E1125 14:01:24.194378 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"102b681c7b99440e1cc3b3aff724d38604f2c3ccd2babbdd6c9531039183cdcd\": container with ID starting with 102b681c7b99440e1cc3b3aff724d38604f2c3ccd2babbdd6c9531039183cdcd not found: ID does not exist" containerID="102b681c7b99440e1cc3b3aff724d38604f2c3ccd2babbdd6c9531039183cdcd" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.194431 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"102b681c7b99440e1cc3b3aff724d38604f2c3ccd2babbdd6c9531039183cdcd"} err="failed to get container status \"102b681c7b99440e1cc3b3aff724d38604f2c3ccd2babbdd6c9531039183cdcd\": rpc error: code = NotFound desc = could not find container \"102b681c7b99440e1cc3b3aff724d38604f2c3ccd2babbdd6c9531039183cdcd\": container with ID starting with 102b681c7b99440e1cc3b3aff724d38604f2c3ccd2babbdd6c9531039183cdcd not found: ID does not exist" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.194467 4702 scope.go:117] "RemoveContainer" containerID="2b311669ef89fab371689904adeadeb2d5759ea31eec30396079cad98594e141" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.197372 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sj8db\" (UniqueName: \"kubernetes.io/projected/815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a-kube-api-access-sj8db\") on node \"crc\" DevicePath \"\"" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.197407 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 14:01:24 crc kubenswrapper[4702]: E1125 14:01:24.197421 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b311669ef89fab371689904adeadeb2d5759ea31eec30396079cad98594e141\": container with ID starting with 2b311669ef89fab371689904adeadeb2d5759ea31eec30396079cad98594e141 not found: ID does not exist" containerID="2b311669ef89fab371689904adeadeb2d5759ea31eec30396079cad98594e141" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.197458 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b311669ef89fab371689904adeadeb2d5759ea31eec30396079cad98594e141"} err="failed to get container status \"2b311669ef89fab371689904adeadeb2d5759ea31eec30396079cad98594e141\": rpc error: code = NotFound desc = could not find container \"2b311669ef89fab371689904adeadeb2d5759ea31eec30396079cad98594e141\": container with ID starting with 2b311669ef89fab371689904adeadeb2d5759ea31eec30396079cad98594e141 not found: ID does not exist" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.197489 4702 scope.go:117] "RemoveContainer" containerID="7a0cb0aac2f05c0b94b27450cab3f1b77c6ca3823e0ccb9f0996de63481ed671" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.212984 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a" (UID: "815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.222316 4702 scope.go:117] "RemoveContainer" containerID="ffddeb21b7ccf95c824e2e11378d2bfdea4f118d2fbca943d28503ef70669bf9" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.247665 4702 scope.go:117] "RemoveContainer" containerID="d4090e4bf5ce943ba63c157eaf77facb57583af02e12a89816809b15970d39a8" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.284434 4702 scope.go:117] "RemoveContainer" containerID="7a0cb0aac2f05c0b94b27450cab3f1b77c6ca3823e0ccb9f0996de63481ed671" Nov 25 14:01:24 crc kubenswrapper[4702]: E1125 14:01:24.285155 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a0cb0aac2f05c0b94b27450cab3f1b77c6ca3823e0ccb9f0996de63481ed671\": container with ID starting with 7a0cb0aac2f05c0b94b27450cab3f1b77c6ca3823e0ccb9f0996de63481ed671 not found: ID does not exist" containerID="7a0cb0aac2f05c0b94b27450cab3f1b77c6ca3823e0ccb9f0996de63481ed671" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.285229 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a0cb0aac2f05c0b94b27450cab3f1b77c6ca3823e0ccb9f0996de63481ed671"} err="failed to get container status \"7a0cb0aac2f05c0b94b27450cab3f1b77c6ca3823e0ccb9f0996de63481ed671\": rpc error: code = NotFound desc = could not find container \"7a0cb0aac2f05c0b94b27450cab3f1b77c6ca3823e0ccb9f0996de63481ed671\": container with ID starting with 7a0cb0aac2f05c0b94b27450cab3f1b77c6ca3823e0ccb9f0996de63481ed671 not found: ID does not exist" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.285294 4702 scope.go:117] "RemoveContainer" containerID="ffddeb21b7ccf95c824e2e11378d2bfdea4f118d2fbca943d28503ef70669bf9" Nov 25 14:01:24 crc kubenswrapper[4702]: E1125 14:01:24.286189 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffddeb21b7ccf95c824e2e11378d2bfdea4f118d2fbca943d28503ef70669bf9\": container with ID starting with ffddeb21b7ccf95c824e2e11378d2bfdea4f118d2fbca943d28503ef70669bf9 not found: ID does not exist" containerID="ffddeb21b7ccf95c824e2e11378d2bfdea4f118d2fbca943d28503ef70669bf9" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.286227 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffddeb21b7ccf95c824e2e11378d2bfdea4f118d2fbca943d28503ef70669bf9"} err="failed to get container status \"ffddeb21b7ccf95c824e2e11378d2bfdea4f118d2fbca943d28503ef70669bf9\": rpc error: code = NotFound desc = could not find container \"ffddeb21b7ccf95c824e2e11378d2bfdea4f118d2fbca943d28503ef70669bf9\": container with ID starting with ffddeb21b7ccf95c824e2e11378d2bfdea4f118d2fbca943d28503ef70669bf9 not found: ID does not exist" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.286265 4702 scope.go:117] "RemoveContainer" containerID="d4090e4bf5ce943ba63c157eaf77facb57583af02e12a89816809b15970d39a8" Nov 25 14:01:24 crc kubenswrapper[4702]: E1125 14:01:24.288442 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4090e4bf5ce943ba63c157eaf77facb57583af02e12a89816809b15970d39a8\": container with ID starting with d4090e4bf5ce943ba63c157eaf77facb57583af02e12a89816809b15970d39a8 not found: ID does not exist" containerID="d4090e4bf5ce943ba63c157eaf77facb57583af02e12a89816809b15970d39a8" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.288471 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4090e4bf5ce943ba63c157eaf77facb57583af02e12a89816809b15970d39a8"} err="failed to get container status \"d4090e4bf5ce943ba63c157eaf77facb57583af02e12a89816809b15970d39a8\": rpc error: code = NotFound desc = could not find container \"d4090e4bf5ce943ba63c157eaf77facb57583af02e12a89816809b15970d39a8\": container with ID starting with d4090e4bf5ce943ba63c157eaf77facb57583af02e12a89816809b15970d39a8 not found: ID does not exist" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.298674 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.391459 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.400353 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/aeb59e3f-26d5-426a-a371-cdc4efd75add-bundle\") pod \"aeb59e3f-26d5-426a-a371-cdc4efd75add\" (UID: \"aeb59e3f-26d5-426a-a371-cdc4efd75add\") " Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.400716 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/aeb59e3f-26d5-426a-a371-cdc4efd75add-util\") pod \"aeb59e3f-26d5-426a-a371-cdc4efd75add\" (UID: \"aeb59e3f-26d5-426a-a371-cdc4efd75add\") " Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.400821 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjm8x\" (UniqueName: \"kubernetes.io/projected/aeb59e3f-26d5-426a-a371-cdc4efd75add-kube-api-access-hjm8x\") pod \"aeb59e3f-26d5-426a-a371-cdc4efd75add\" (UID: \"aeb59e3f-26d5-426a-a371-cdc4efd75add\") " Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.401498 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aeb59e3f-26d5-426a-a371-cdc4efd75add-bundle" (OuterVolumeSpecName: "bundle") pod "aeb59e3f-26d5-426a-a371-cdc4efd75add" (UID: "aeb59e3f-26d5-426a-a371-cdc4efd75add"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.406085 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aeb59e3f-26d5-426a-a371-cdc4efd75add-kube-api-access-hjm8x" (OuterVolumeSpecName: "kube-api-access-hjm8x") pod "aeb59e3f-26d5-426a-a371-cdc4efd75add" (UID: "aeb59e3f-26d5-426a-a371-cdc4efd75add"). InnerVolumeSpecName "kube-api-access-hjm8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.409835 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vkrfk"] Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.414042 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vkrfk"] Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.416183 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aeb59e3f-26d5-426a-a371-cdc4efd75add-util" (OuterVolumeSpecName: "util") pod "aeb59e3f-26d5-426a-a371-cdc4efd75add" (UID: "aeb59e3f-26d5-426a-a371-cdc4efd75add"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.502040 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjm8x\" (UniqueName: \"kubernetes.io/projected/aeb59e3f-26d5-426a-a371-cdc4efd75add-kube-api-access-hjm8x\") on node \"crc\" DevicePath \"\"" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.502084 4702 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/aeb59e3f-26d5-426a-a371-cdc4efd75add-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 14:01:24 crc kubenswrapper[4702]: I1125 14:01:24.502095 4702 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/aeb59e3f-26d5-426a-a371-cdc4efd75add-util\") on node \"crc\" DevicePath \"\"" Nov 25 14:01:25 crc kubenswrapper[4702]: I1125 14:01:25.059149 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5" event={"ID":"aeb59e3f-26d5-426a-a371-cdc4efd75add","Type":"ContainerDied","Data":"0c323de7ce63c0fbdc7978a4bb8416a3d684ce193839ba035d396d2a4ebd688f"} Nov 25 14:01:25 crc kubenswrapper[4702]: I1125 14:01:25.059650 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c323de7ce63c0fbdc7978a4bb8416a3d684ce193839ba035d396d2a4ebd688f" Nov 25 14:01:25 crc kubenswrapper[4702]: I1125 14:01:25.059344 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5" Nov 25 14:01:25 crc kubenswrapper[4702]: I1125 14:01:25.837521 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a7e12c8-26bc-4d1d-877c-88b89091b78e" path="/var/lib/kubelet/pods/1a7e12c8-26bc-4d1d-877c-88b89091b78e/volumes" Nov 25 14:01:25 crc kubenswrapper[4702]: I1125 14:01:25.838946 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a" path="/var/lib/kubelet/pods/815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a/volumes" Nov 25 14:01:26 crc kubenswrapper[4702]: I1125 14:01:26.853100 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 14:01:26 crc kubenswrapper[4702]: I1125 14:01:26.933379 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 14:01:30 crc kubenswrapper[4702]: I1125 14:01:30.096951 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 14:01:30 crc kubenswrapper[4702]: I1125 14:01:30.187626 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 14:01:39 crc kubenswrapper[4702]: I1125 14:01:39.113278 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-nzffd"] Nov 25 14:01:39 crc kubenswrapper[4702]: E1125 14:01:39.114336 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a7e12c8-26bc-4d1d-877c-88b89091b78e" containerName="extract-utilities" Nov 25 14:01:39 crc kubenswrapper[4702]: I1125 14:01:39.114351 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a7e12c8-26bc-4d1d-877c-88b89091b78e" containerName="extract-utilities" Nov 25 14:01:39 crc kubenswrapper[4702]: E1125 14:01:39.114377 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a7e12c8-26bc-4d1d-877c-88b89091b78e" containerName="extract-content" Nov 25 14:01:39 crc kubenswrapper[4702]: I1125 14:01:39.114384 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a7e12c8-26bc-4d1d-877c-88b89091b78e" containerName="extract-content" Nov 25 14:01:39 crc kubenswrapper[4702]: E1125 14:01:39.114395 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeb59e3f-26d5-426a-a371-cdc4efd75add" containerName="util" Nov 25 14:01:39 crc kubenswrapper[4702]: I1125 14:01:39.114402 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeb59e3f-26d5-426a-a371-cdc4efd75add" containerName="util" Nov 25 14:01:39 crc kubenswrapper[4702]: E1125 14:01:39.114413 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a7e12c8-26bc-4d1d-877c-88b89091b78e" containerName="registry-server" Nov 25 14:01:39 crc kubenswrapper[4702]: I1125 14:01:39.114418 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a7e12c8-26bc-4d1d-877c-88b89091b78e" containerName="registry-server" Nov 25 14:01:39 crc kubenswrapper[4702]: E1125 14:01:39.114426 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a" containerName="registry-server" Nov 25 14:01:39 crc kubenswrapper[4702]: I1125 14:01:39.114432 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a" containerName="registry-server" Nov 25 14:01:39 crc kubenswrapper[4702]: E1125 14:01:39.114441 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a" containerName="extract-utilities" Nov 25 14:01:39 crc kubenswrapper[4702]: I1125 14:01:39.114447 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a" containerName="extract-utilities" Nov 25 14:01:39 crc kubenswrapper[4702]: E1125 14:01:39.114460 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeb59e3f-26d5-426a-a371-cdc4efd75add" containerName="pull" Nov 25 14:01:39 crc kubenswrapper[4702]: I1125 14:01:39.114466 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeb59e3f-26d5-426a-a371-cdc4efd75add" containerName="pull" Nov 25 14:01:39 crc kubenswrapper[4702]: E1125 14:01:39.114474 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeb59e3f-26d5-426a-a371-cdc4efd75add" containerName="extract" Nov 25 14:01:39 crc kubenswrapper[4702]: I1125 14:01:39.114480 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeb59e3f-26d5-426a-a371-cdc4efd75add" containerName="extract" Nov 25 14:01:39 crc kubenswrapper[4702]: E1125 14:01:39.114486 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a" containerName="extract-content" Nov 25 14:01:39 crc kubenswrapper[4702]: I1125 14:01:39.114491 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a" containerName="extract-content" Nov 25 14:01:39 crc kubenswrapper[4702]: I1125 14:01:39.114591 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a7e12c8-26bc-4d1d-877c-88b89091b78e" containerName="registry-server" Nov 25 14:01:39 crc kubenswrapper[4702]: I1125 14:01:39.114605 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="aeb59e3f-26d5-426a-a371-cdc4efd75add" containerName="extract" Nov 25 14:01:39 crc kubenswrapper[4702]: I1125 14:01:39.114616 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="815f3b6c-d4bd-4bf7-92f7-cd3af4b1e85a" containerName="registry-server" Nov 25 14:01:39 crc kubenswrapper[4702]: I1125 14:01:39.115079 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-nzffd" Nov 25 14:01:39 crc kubenswrapper[4702]: I1125 14:01:39.118003 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-dockercfg-7w7tz" Nov 25 14:01:39 crc kubenswrapper[4702]: I1125 14:01:39.130444 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-nzffd"] Nov 25 14:01:39 crc kubenswrapper[4702]: I1125 14:01:39.267082 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t5l8\" (UniqueName: \"kubernetes.io/projected/3106c490-5a52-4dd1-9ef8-9ee2325a7d62-kube-api-access-9t5l8\") pod \"rabbitmq-cluster-operator-779fc9694b-nzffd\" (UID: \"3106c490-5a52-4dd1-9ef8-9ee2325a7d62\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-nzffd" Nov 25 14:01:39 crc kubenswrapper[4702]: I1125 14:01:39.369371 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9t5l8\" (UniqueName: \"kubernetes.io/projected/3106c490-5a52-4dd1-9ef8-9ee2325a7d62-kube-api-access-9t5l8\") pod \"rabbitmq-cluster-operator-779fc9694b-nzffd\" (UID: \"3106c490-5a52-4dd1-9ef8-9ee2325a7d62\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-nzffd" Nov 25 14:01:39 crc kubenswrapper[4702]: I1125 14:01:39.391540 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9t5l8\" (UniqueName: \"kubernetes.io/projected/3106c490-5a52-4dd1-9ef8-9ee2325a7d62-kube-api-access-9t5l8\") pod \"rabbitmq-cluster-operator-779fc9694b-nzffd\" (UID: \"3106c490-5a52-4dd1-9ef8-9ee2325a7d62\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-nzffd" Nov 25 14:01:39 crc kubenswrapper[4702]: I1125 14:01:39.432092 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-nzffd" Nov 25 14:01:39 crc kubenswrapper[4702]: I1125 14:01:39.865706 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-nzffd"] Nov 25 14:01:40 crc kubenswrapper[4702]: I1125 14:01:40.162488 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-nzffd" event={"ID":"3106c490-5a52-4dd1-9ef8-9ee2325a7d62","Type":"ContainerStarted","Data":"77ba8be3f04adb8a42a0f905078bc1dcb63c672b1bb5568f14169839ca17f79d"} Nov 25 14:01:44 crc kubenswrapper[4702]: I1125 14:01:44.194338 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-nzffd" event={"ID":"3106c490-5a52-4dd1-9ef8-9ee2325a7d62","Type":"ContainerStarted","Data":"db95164e578b3f445d1fcadd2fe0741754e05a14af2e50193361777af87abb93"} Nov 25 14:01:44 crc kubenswrapper[4702]: I1125 14:01:44.214390 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-nzffd" podStartSLOduration=1.561488478 podStartE2EDuration="5.214357333s" podCreationTimestamp="2025-11-25 14:01:39 +0000 UTC" firstStartedPulling="2025-11-25 14:01:39.880896417 +0000 UTC m=+909.043546867" lastFinishedPulling="2025-11-25 14:01:43.533765262 +0000 UTC m=+912.696415722" observedRunningTime="2025-11-25 14:01:44.213355857 +0000 UTC m=+913.376006307" watchObservedRunningTime="2025-11-25 14:01:44.214357333 +0000 UTC m=+913.377007803" Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.473933 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.475671 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.478171 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-erlang-cookie" Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.478536 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-plugins-conf" Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.478611 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-default-user" Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.483266 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-server-conf" Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.483497 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-server-dockercfg-clh2p" Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.495749 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.667865 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3454fb60-87b9-45de-b8b2-13548d88cb1e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3454fb60-87b9-45de-b8b2-13548d88cb1e\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.667924 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3454fb60-87b9-45de-b8b2-13548d88cb1e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3454fb60-87b9-45de-b8b2-13548d88cb1e\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.667959 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3454fb60-87b9-45de-b8b2-13548d88cb1e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3454fb60-87b9-45de-b8b2-13548d88cb1e\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.667975 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3454fb60-87b9-45de-b8b2-13548d88cb1e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3454fb60-87b9-45de-b8b2-13548d88cb1e\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.667992 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3454fb60-87b9-45de-b8b2-13548d88cb1e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3454fb60-87b9-45de-b8b2-13548d88cb1e\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.668056 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3454fb60-87b9-45de-b8b2-13548d88cb1e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3454fb60-87b9-45de-b8b2-13548d88cb1e\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.668088 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ee2f5863-633c-4f23-b012-dd96bbed8190\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ee2f5863-633c-4f23-b012-dd96bbed8190\") pod \"rabbitmq-server-0\" (UID: \"3454fb60-87b9-45de-b8b2-13548d88cb1e\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.668111 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmd6j\" (UniqueName: \"kubernetes.io/projected/3454fb60-87b9-45de-b8b2-13548d88cb1e-kube-api-access-vmd6j\") pod \"rabbitmq-server-0\" (UID: \"3454fb60-87b9-45de-b8b2-13548d88cb1e\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.770309 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3454fb60-87b9-45de-b8b2-13548d88cb1e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3454fb60-87b9-45de-b8b2-13548d88cb1e\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.770403 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3454fb60-87b9-45de-b8b2-13548d88cb1e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3454fb60-87b9-45de-b8b2-13548d88cb1e\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.770447 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3454fb60-87b9-45de-b8b2-13548d88cb1e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3454fb60-87b9-45de-b8b2-13548d88cb1e\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.770470 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3454fb60-87b9-45de-b8b2-13548d88cb1e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3454fb60-87b9-45de-b8b2-13548d88cb1e\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.770494 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3454fb60-87b9-45de-b8b2-13548d88cb1e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3454fb60-87b9-45de-b8b2-13548d88cb1e\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.770573 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3454fb60-87b9-45de-b8b2-13548d88cb1e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3454fb60-87b9-45de-b8b2-13548d88cb1e\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.770609 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ee2f5863-633c-4f23-b012-dd96bbed8190\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ee2f5863-633c-4f23-b012-dd96bbed8190\") pod \"rabbitmq-server-0\" (UID: \"3454fb60-87b9-45de-b8b2-13548d88cb1e\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.770636 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmd6j\" (UniqueName: \"kubernetes.io/projected/3454fb60-87b9-45de-b8b2-13548d88cb1e-kube-api-access-vmd6j\") pod \"rabbitmq-server-0\" (UID: \"3454fb60-87b9-45de-b8b2-13548d88cb1e\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.771443 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3454fb60-87b9-45de-b8b2-13548d88cb1e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3454fb60-87b9-45de-b8b2-13548d88cb1e\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.771980 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3454fb60-87b9-45de-b8b2-13548d88cb1e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3454fb60-87b9-45de-b8b2-13548d88cb1e\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.772883 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3454fb60-87b9-45de-b8b2-13548d88cb1e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3454fb60-87b9-45de-b8b2-13548d88cb1e\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.777137 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3454fb60-87b9-45de-b8b2-13548d88cb1e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3454fb60-87b9-45de-b8b2-13548d88cb1e\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.777799 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3454fb60-87b9-45de-b8b2-13548d88cb1e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3454fb60-87b9-45de-b8b2-13548d88cb1e\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.786709 4702 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.786991 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ee2f5863-633c-4f23-b012-dd96bbed8190\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ee2f5863-633c-4f23-b012-dd96bbed8190\") pod \"rabbitmq-server-0\" (UID: \"3454fb60-87b9-45de-b8b2-13548d88cb1e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1e8e3a05721d6e45b807327ae43ebb6522a9ae0c39fdeddb16d2bff82c3b4b1c/globalmount\"" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.789774 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmd6j\" (UniqueName: \"kubernetes.io/projected/3454fb60-87b9-45de-b8b2-13548d88cb1e-kube-api-access-vmd6j\") pod \"rabbitmq-server-0\" (UID: \"3454fb60-87b9-45de-b8b2-13548d88cb1e\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.792631 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3454fb60-87b9-45de-b8b2-13548d88cb1e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3454fb60-87b9-45de-b8b2-13548d88cb1e\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 14:01:48 crc kubenswrapper[4702]: I1125 14:01:48.825035 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ee2f5863-633c-4f23-b012-dd96bbed8190\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ee2f5863-633c-4f23-b012-dd96bbed8190\") pod \"rabbitmq-server-0\" (UID: \"3454fb60-87b9-45de-b8b2-13548d88cb1e\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 14:01:49 crc kubenswrapper[4702]: I1125 14:01:49.096384 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 14:01:49 crc kubenswrapper[4702]: I1125 14:01:49.360955 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 25 14:01:50 crc kubenswrapper[4702]: I1125 14:01:50.002622 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-v8qqn"] Nov 25 14:01:50 crc kubenswrapper[4702]: I1125 14:01:50.003770 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-v8qqn" Nov 25 14:01:50 crc kubenswrapper[4702]: I1125 14:01:50.006370 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-index-dockercfg-gxk7n" Nov 25 14:01:50 crc kubenswrapper[4702]: I1125 14:01:50.018670 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-v8qqn"] Nov 25 14:01:50 crc kubenswrapper[4702]: I1125 14:01:50.196690 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cn264\" (UniqueName: \"kubernetes.io/projected/3969eead-a612-481f-ac52-a2b91a902fea-kube-api-access-cn264\") pod \"keystone-operator-index-v8qqn\" (UID: \"3969eead-a612-481f-ac52-a2b91a902fea\") " pod="openstack-operators/keystone-operator-index-v8qqn" Nov 25 14:01:50 crc kubenswrapper[4702]: I1125 14:01:50.279328 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"3454fb60-87b9-45de-b8b2-13548d88cb1e","Type":"ContainerStarted","Data":"ca949413b4b6ea209ad6ad0c394e9a245551fe295b53fd0ac0450dd5536c4ace"} Nov 25 14:01:50 crc kubenswrapper[4702]: I1125 14:01:50.298596 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cn264\" (UniqueName: \"kubernetes.io/projected/3969eead-a612-481f-ac52-a2b91a902fea-kube-api-access-cn264\") pod \"keystone-operator-index-v8qqn\" (UID: \"3969eead-a612-481f-ac52-a2b91a902fea\") " pod="openstack-operators/keystone-operator-index-v8qqn" Nov 25 14:01:50 crc kubenswrapper[4702]: I1125 14:01:50.328640 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cn264\" (UniqueName: \"kubernetes.io/projected/3969eead-a612-481f-ac52-a2b91a902fea-kube-api-access-cn264\") pod \"keystone-operator-index-v8qqn\" (UID: \"3969eead-a612-481f-ac52-a2b91a902fea\") " pod="openstack-operators/keystone-operator-index-v8qqn" Nov 25 14:01:50 crc kubenswrapper[4702]: I1125 14:01:50.624800 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-v8qqn" Nov 25 14:01:51 crc kubenswrapper[4702]: I1125 14:01:51.077540 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-v8qqn"] Nov 25 14:01:51 crc kubenswrapper[4702]: W1125 14:01:51.085986 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3969eead_a612_481f_ac52_a2b91a902fea.slice/crio-8da6e6ab23340ed42c91dc20de1d66b6a9b12e66f2e29421875b978113de1639 WatchSource:0}: Error finding container 8da6e6ab23340ed42c91dc20de1d66b6a9b12e66f2e29421875b978113de1639: Status 404 returned error can't find the container with id 8da6e6ab23340ed42c91dc20de1d66b6a9b12e66f2e29421875b978113de1639 Nov 25 14:01:51 crc kubenswrapper[4702]: I1125 14:01:51.289280 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-v8qqn" event={"ID":"3969eead-a612-481f-ac52-a2b91a902fea","Type":"ContainerStarted","Data":"8da6e6ab23340ed42c91dc20de1d66b6a9b12e66f2e29421875b978113de1639"} Nov 25 14:01:54 crc kubenswrapper[4702]: I1125 14:01:54.391822 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-v8qqn"] Nov 25 14:01:55 crc kubenswrapper[4702]: I1125 14:01:55.006140 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-rx4hj"] Nov 25 14:01:55 crc kubenswrapper[4702]: I1125 14:01:55.007209 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-rx4hj" Nov 25 14:01:55 crc kubenswrapper[4702]: I1125 14:01:55.018793 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-rx4hj"] Nov 25 14:01:55 crc kubenswrapper[4702]: I1125 14:01:55.087608 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jl2s4\" (UniqueName: \"kubernetes.io/projected/ca7f19bd-2112-4f04-bd40-c8017f61804d-kube-api-access-jl2s4\") pod \"keystone-operator-index-rx4hj\" (UID: \"ca7f19bd-2112-4f04-bd40-c8017f61804d\") " pod="openstack-operators/keystone-operator-index-rx4hj" Nov 25 14:01:55 crc kubenswrapper[4702]: I1125 14:01:55.188834 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jl2s4\" (UniqueName: \"kubernetes.io/projected/ca7f19bd-2112-4f04-bd40-c8017f61804d-kube-api-access-jl2s4\") pod \"keystone-operator-index-rx4hj\" (UID: \"ca7f19bd-2112-4f04-bd40-c8017f61804d\") " pod="openstack-operators/keystone-operator-index-rx4hj" Nov 25 14:01:55 crc kubenswrapper[4702]: I1125 14:01:55.212859 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jl2s4\" (UniqueName: \"kubernetes.io/projected/ca7f19bd-2112-4f04-bd40-c8017f61804d-kube-api-access-jl2s4\") pod \"keystone-operator-index-rx4hj\" (UID: \"ca7f19bd-2112-4f04-bd40-c8017f61804d\") " pod="openstack-operators/keystone-operator-index-rx4hj" Nov 25 14:01:55 crc kubenswrapper[4702]: I1125 14:01:55.375998 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-rx4hj" Nov 25 14:01:57 crc kubenswrapper[4702]: I1125 14:01:57.396648 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-rx4hj"] Nov 25 14:01:57 crc kubenswrapper[4702]: W1125 14:01:57.406037 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca7f19bd_2112_4f04_bd40_c8017f61804d.slice/crio-2dc61f323bc373e57f0ddf294d7252febacdddf293f748b726fe13f3bf1519f5 WatchSource:0}: Error finding container 2dc61f323bc373e57f0ddf294d7252febacdddf293f748b726fe13f3bf1519f5: Status 404 returned error can't find the container with id 2dc61f323bc373e57f0ddf294d7252febacdddf293f748b726fe13f3bf1519f5 Nov 25 14:01:58 crc kubenswrapper[4702]: I1125 14:01:58.348499 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-v8qqn" event={"ID":"3969eead-a612-481f-ac52-a2b91a902fea","Type":"ContainerStarted","Data":"669247e2de0afa4dd687eeead7ff230da7bcb3cd15efc30e42bde1884799cc2e"} Nov 25 14:01:58 crc kubenswrapper[4702]: I1125 14:01:58.348696 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/keystone-operator-index-v8qqn" podUID="3969eead-a612-481f-ac52-a2b91a902fea" containerName="registry-server" containerID="cri-o://669247e2de0afa4dd687eeead7ff230da7bcb3cd15efc30e42bde1884799cc2e" gracePeriod=2 Nov 25 14:01:58 crc kubenswrapper[4702]: I1125 14:01:58.350520 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-rx4hj" event={"ID":"ca7f19bd-2112-4f04-bd40-c8017f61804d","Type":"ContainerStarted","Data":"b9d57f052ba80e14e99056c1a4511bdb085dbf3fea90d2b064718816fb0a6a5b"} Nov 25 14:01:58 crc kubenswrapper[4702]: I1125 14:01:58.350552 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-rx4hj" event={"ID":"ca7f19bd-2112-4f04-bd40-c8017f61804d","Type":"ContainerStarted","Data":"2dc61f323bc373e57f0ddf294d7252febacdddf293f748b726fe13f3bf1519f5"} Nov 25 14:01:58 crc kubenswrapper[4702]: I1125 14:01:58.368162 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-v8qqn" podStartSLOduration=4.141980606 podStartE2EDuration="9.368132806s" podCreationTimestamp="2025-11-25 14:01:49 +0000 UTC" firstStartedPulling="2025-11-25 14:01:51.089010172 +0000 UTC m=+920.251660622" lastFinishedPulling="2025-11-25 14:01:56.315162372 +0000 UTC m=+925.477812822" observedRunningTime="2025-11-25 14:01:58.36788051 +0000 UTC m=+927.530530960" watchObservedRunningTime="2025-11-25 14:01:58.368132806 +0000 UTC m=+927.530783256" Nov 25 14:01:58 crc kubenswrapper[4702]: I1125 14:01:58.999369 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-v8qqn" Nov 25 14:01:59 crc kubenswrapper[4702]: I1125 14:01:59.017870 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-rx4hj" podStartSLOduration=4.422350264 podStartE2EDuration="5.017838803s" podCreationTimestamp="2025-11-25 14:01:54 +0000 UTC" firstStartedPulling="2025-11-25 14:01:57.411966501 +0000 UTC m=+926.574616961" lastFinishedPulling="2025-11-25 14:01:58.00745505 +0000 UTC m=+927.170105500" observedRunningTime="2025-11-25 14:01:58.394020894 +0000 UTC m=+927.556671344" watchObservedRunningTime="2025-11-25 14:01:59.017838803 +0000 UTC m=+928.180489253" Nov 25 14:01:59 crc kubenswrapper[4702]: I1125 14:01:59.192349 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cn264\" (UniqueName: \"kubernetes.io/projected/3969eead-a612-481f-ac52-a2b91a902fea-kube-api-access-cn264\") pod \"3969eead-a612-481f-ac52-a2b91a902fea\" (UID: \"3969eead-a612-481f-ac52-a2b91a902fea\") " Nov 25 14:01:59 crc kubenswrapper[4702]: I1125 14:01:59.200664 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3969eead-a612-481f-ac52-a2b91a902fea-kube-api-access-cn264" (OuterVolumeSpecName: "kube-api-access-cn264") pod "3969eead-a612-481f-ac52-a2b91a902fea" (UID: "3969eead-a612-481f-ac52-a2b91a902fea"). InnerVolumeSpecName "kube-api-access-cn264". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:01:59 crc kubenswrapper[4702]: I1125 14:01:59.294184 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cn264\" (UniqueName: \"kubernetes.io/projected/3969eead-a612-481f-ac52-a2b91a902fea-kube-api-access-cn264\") on node \"crc\" DevicePath \"\"" Nov 25 14:01:59 crc kubenswrapper[4702]: I1125 14:01:59.361201 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"3454fb60-87b9-45de-b8b2-13548d88cb1e","Type":"ContainerStarted","Data":"770cda36d12cd5d01cc54f5e98b263ef301e050bba16c09a4a3a2b4ca3be2a59"} Nov 25 14:01:59 crc kubenswrapper[4702]: I1125 14:01:59.363430 4702 generic.go:334] "Generic (PLEG): container finished" podID="3969eead-a612-481f-ac52-a2b91a902fea" containerID="669247e2de0afa4dd687eeead7ff230da7bcb3cd15efc30e42bde1884799cc2e" exitCode=0 Nov 25 14:01:59 crc kubenswrapper[4702]: I1125 14:01:59.363493 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-v8qqn" Nov 25 14:01:59 crc kubenswrapper[4702]: I1125 14:01:59.363563 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-v8qqn" event={"ID":"3969eead-a612-481f-ac52-a2b91a902fea","Type":"ContainerDied","Data":"669247e2de0afa4dd687eeead7ff230da7bcb3cd15efc30e42bde1884799cc2e"} Nov 25 14:01:59 crc kubenswrapper[4702]: I1125 14:01:59.363641 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-v8qqn" event={"ID":"3969eead-a612-481f-ac52-a2b91a902fea","Type":"ContainerDied","Data":"8da6e6ab23340ed42c91dc20de1d66b6a9b12e66f2e29421875b978113de1639"} Nov 25 14:01:59 crc kubenswrapper[4702]: I1125 14:01:59.363666 4702 scope.go:117] "RemoveContainer" containerID="669247e2de0afa4dd687eeead7ff230da7bcb3cd15efc30e42bde1884799cc2e" Nov 25 14:01:59 crc kubenswrapper[4702]: I1125 14:01:59.394527 4702 scope.go:117] "RemoveContainer" containerID="669247e2de0afa4dd687eeead7ff230da7bcb3cd15efc30e42bde1884799cc2e" Nov 25 14:01:59 crc kubenswrapper[4702]: E1125 14:01:59.395209 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"669247e2de0afa4dd687eeead7ff230da7bcb3cd15efc30e42bde1884799cc2e\": container with ID starting with 669247e2de0afa4dd687eeead7ff230da7bcb3cd15efc30e42bde1884799cc2e not found: ID does not exist" containerID="669247e2de0afa4dd687eeead7ff230da7bcb3cd15efc30e42bde1884799cc2e" Nov 25 14:01:59 crc kubenswrapper[4702]: I1125 14:01:59.395286 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"669247e2de0afa4dd687eeead7ff230da7bcb3cd15efc30e42bde1884799cc2e"} err="failed to get container status \"669247e2de0afa4dd687eeead7ff230da7bcb3cd15efc30e42bde1884799cc2e\": rpc error: code = NotFound desc = could not find container \"669247e2de0afa4dd687eeead7ff230da7bcb3cd15efc30e42bde1884799cc2e\": container with ID starting with 669247e2de0afa4dd687eeead7ff230da7bcb3cd15efc30e42bde1884799cc2e not found: ID does not exist" Nov 25 14:01:59 crc kubenswrapper[4702]: I1125 14:01:59.402933 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-v8qqn"] Nov 25 14:01:59 crc kubenswrapper[4702]: I1125 14:01:59.409921 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/keystone-operator-index-v8qqn"] Nov 25 14:01:59 crc kubenswrapper[4702]: I1125 14:01:59.811315 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3969eead-a612-481f-ac52-a2b91a902fea" path="/var/lib/kubelet/pods/3969eead-a612-481f-ac52-a2b91a902fea/volumes" Nov 25 14:02:05 crc kubenswrapper[4702]: I1125 14:02:05.376566 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/keystone-operator-index-rx4hj" Nov 25 14:02:05 crc kubenswrapper[4702]: I1125 14:02:05.377367 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-index-rx4hj" Nov 25 14:02:05 crc kubenswrapper[4702]: I1125 14:02:05.415681 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/keystone-operator-index-rx4hj" Nov 25 14:02:06 crc kubenswrapper[4702]: I1125 14:02:06.450686 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-index-rx4hj" Nov 25 14:02:15 crc kubenswrapper[4702]: I1125 14:02:15.039939 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd"] Nov 25 14:02:15 crc kubenswrapper[4702]: E1125 14:02:15.040792 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3969eead-a612-481f-ac52-a2b91a902fea" containerName="registry-server" Nov 25 14:02:15 crc kubenswrapper[4702]: I1125 14:02:15.040807 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="3969eead-a612-481f-ac52-a2b91a902fea" containerName="registry-server" Nov 25 14:02:15 crc kubenswrapper[4702]: I1125 14:02:15.040954 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="3969eead-a612-481f-ac52-a2b91a902fea" containerName="registry-server" Nov 25 14:02:15 crc kubenswrapper[4702]: I1125 14:02:15.042045 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd" Nov 25 14:02:15 crc kubenswrapper[4702]: I1125 14:02:15.045272 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-6r4jq" Nov 25 14:02:15 crc kubenswrapper[4702]: I1125 14:02:15.055351 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd"] Nov 25 14:02:15 crc kubenswrapper[4702]: I1125 14:02:15.152813 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lpww\" (UniqueName: \"kubernetes.io/projected/0f4314a6-9010-4789-816c-3bc52836c02e-kube-api-access-9lpww\") pod \"62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd\" (UID: \"0f4314a6-9010-4789-816c-3bc52836c02e\") " pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd" Nov 25 14:02:15 crc kubenswrapper[4702]: I1125 14:02:15.152947 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0f4314a6-9010-4789-816c-3bc52836c02e-bundle\") pod \"62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd\" (UID: \"0f4314a6-9010-4789-816c-3bc52836c02e\") " pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd" Nov 25 14:02:15 crc kubenswrapper[4702]: I1125 14:02:15.152977 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0f4314a6-9010-4789-816c-3bc52836c02e-util\") pod \"62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd\" (UID: \"0f4314a6-9010-4789-816c-3bc52836c02e\") " pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd" Nov 25 14:02:15 crc kubenswrapper[4702]: I1125 14:02:15.254668 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0f4314a6-9010-4789-816c-3bc52836c02e-util\") pod \"62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd\" (UID: \"0f4314a6-9010-4789-816c-3bc52836c02e\") " pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd" Nov 25 14:02:15 crc kubenswrapper[4702]: I1125 14:02:15.254747 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lpww\" (UniqueName: \"kubernetes.io/projected/0f4314a6-9010-4789-816c-3bc52836c02e-kube-api-access-9lpww\") pod \"62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd\" (UID: \"0f4314a6-9010-4789-816c-3bc52836c02e\") " pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd" Nov 25 14:02:15 crc kubenswrapper[4702]: I1125 14:02:15.254819 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0f4314a6-9010-4789-816c-3bc52836c02e-bundle\") pod \"62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd\" (UID: \"0f4314a6-9010-4789-816c-3bc52836c02e\") " pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd" Nov 25 14:02:15 crc kubenswrapper[4702]: I1125 14:02:15.255279 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0f4314a6-9010-4789-816c-3bc52836c02e-bundle\") pod \"62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd\" (UID: \"0f4314a6-9010-4789-816c-3bc52836c02e\") " pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd" Nov 25 14:02:15 crc kubenswrapper[4702]: I1125 14:02:15.255344 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0f4314a6-9010-4789-816c-3bc52836c02e-util\") pod \"62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd\" (UID: \"0f4314a6-9010-4789-816c-3bc52836c02e\") " pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd" Nov 25 14:02:15 crc kubenswrapper[4702]: I1125 14:02:15.283924 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lpww\" (UniqueName: \"kubernetes.io/projected/0f4314a6-9010-4789-816c-3bc52836c02e-kube-api-access-9lpww\") pod \"62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd\" (UID: \"0f4314a6-9010-4789-816c-3bc52836c02e\") " pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd" Nov 25 14:02:15 crc kubenswrapper[4702]: I1125 14:02:15.439980 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd" Nov 25 14:02:15 crc kubenswrapper[4702]: I1125 14:02:15.869266 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd"] Nov 25 14:02:15 crc kubenswrapper[4702]: W1125 14:02:15.873263 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f4314a6_9010_4789_816c_3bc52836c02e.slice/crio-1d2454cdd3804a8e2c52446c10de3cfe35c1bbb7dd75268e5c05a5291b2aea69 WatchSource:0}: Error finding container 1d2454cdd3804a8e2c52446c10de3cfe35c1bbb7dd75268e5c05a5291b2aea69: Status 404 returned error can't find the container with id 1d2454cdd3804a8e2c52446c10de3cfe35c1bbb7dd75268e5c05a5291b2aea69 Nov 25 14:02:16 crc kubenswrapper[4702]: I1125 14:02:16.477368 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd" event={"ID":"0f4314a6-9010-4789-816c-3bc52836c02e","Type":"ContainerDied","Data":"d06ca5e82692edee9b8b168122c44148dda7a3d2bf262fae681112073f3c984e"} Nov 25 14:02:16 crc kubenswrapper[4702]: I1125 14:02:16.477975 4702 generic.go:334] "Generic (PLEG): container finished" podID="0f4314a6-9010-4789-816c-3bc52836c02e" containerID="d06ca5e82692edee9b8b168122c44148dda7a3d2bf262fae681112073f3c984e" exitCode=0 Nov 25 14:02:16 crc kubenswrapper[4702]: I1125 14:02:16.478061 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd" event={"ID":"0f4314a6-9010-4789-816c-3bc52836c02e","Type":"ContainerStarted","Data":"1d2454cdd3804a8e2c52446c10de3cfe35c1bbb7dd75268e5c05a5291b2aea69"} Nov 25 14:02:18 crc kubenswrapper[4702]: I1125 14:02:18.514616 4702 generic.go:334] "Generic (PLEG): container finished" podID="0f4314a6-9010-4789-816c-3bc52836c02e" containerID="02a8f13dfb9a5e4ac1da32215883f104114261d3951bcbaede9463d65fb79680" exitCode=0 Nov 25 14:02:18 crc kubenswrapper[4702]: I1125 14:02:18.515006 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd" event={"ID":"0f4314a6-9010-4789-816c-3bc52836c02e","Type":"ContainerDied","Data":"02a8f13dfb9a5e4ac1da32215883f104114261d3951bcbaede9463d65fb79680"} Nov 25 14:02:19 crc kubenswrapper[4702]: I1125 14:02:19.524883 4702 generic.go:334] "Generic (PLEG): container finished" podID="0f4314a6-9010-4789-816c-3bc52836c02e" containerID="4a6970fe2aea189613490ec7f49a8553b6f73e6a8f50747ab6997b2663f0ce09" exitCode=0 Nov 25 14:02:19 crc kubenswrapper[4702]: I1125 14:02:19.524928 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd" event={"ID":"0f4314a6-9010-4789-816c-3bc52836c02e","Type":"ContainerDied","Data":"4a6970fe2aea189613490ec7f49a8553b6f73e6a8f50747ab6997b2663f0ce09"} Nov 25 14:02:21 crc kubenswrapper[4702]: I1125 14:02:21.689688 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd" Nov 25 14:02:21 crc kubenswrapper[4702]: I1125 14:02:21.767373 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0f4314a6-9010-4789-816c-3bc52836c02e-bundle\") pod \"0f4314a6-9010-4789-816c-3bc52836c02e\" (UID: \"0f4314a6-9010-4789-816c-3bc52836c02e\") " Nov 25 14:02:21 crc kubenswrapper[4702]: I1125 14:02:21.767522 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lpww\" (UniqueName: \"kubernetes.io/projected/0f4314a6-9010-4789-816c-3bc52836c02e-kube-api-access-9lpww\") pod \"0f4314a6-9010-4789-816c-3bc52836c02e\" (UID: \"0f4314a6-9010-4789-816c-3bc52836c02e\") " Nov 25 14:02:21 crc kubenswrapper[4702]: I1125 14:02:21.767679 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0f4314a6-9010-4789-816c-3bc52836c02e-util\") pod \"0f4314a6-9010-4789-816c-3bc52836c02e\" (UID: \"0f4314a6-9010-4789-816c-3bc52836c02e\") " Nov 25 14:02:21 crc kubenswrapper[4702]: I1125 14:02:21.769440 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f4314a6-9010-4789-816c-3bc52836c02e-bundle" (OuterVolumeSpecName: "bundle") pod "0f4314a6-9010-4789-816c-3bc52836c02e" (UID: "0f4314a6-9010-4789-816c-3bc52836c02e"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:02:21 crc kubenswrapper[4702]: I1125 14:02:21.772574 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f4314a6-9010-4789-816c-3bc52836c02e-kube-api-access-9lpww" (OuterVolumeSpecName: "kube-api-access-9lpww") pod "0f4314a6-9010-4789-816c-3bc52836c02e" (UID: "0f4314a6-9010-4789-816c-3bc52836c02e"). InnerVolumeSpecName "kube-api-access-9lpww". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:02:21 crc kubenswrapper[4702]: I1125 14:02:21.868878 4702 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0f4314a6-9010-4789-816c-3bc52836c02e-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 14:02:21 crc kubenswrapper[4702]: I1125 14:02:21.868916 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lpww\" (UniqueName: \"kubernetes.io/projected/0f4314a6-9010-4789-816c-3bc52836c02e-kube-api-access-9lpww\") on node \"crc\" DevicePath \"\"" Nov 25 14:02:21 crc kubenswrapper[4702]: I1125 14:02:21.873225 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f4314a6-9010-4789-816c-3bc52836c02e-util" (OuterVolumeSpecName: "util") pod "0f4314a6-9010-4789-816c-3bc52836c02e" (UID: "0f4314a6-9010-4789-816c-3bc52836c02e"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:02:21 crc kubenswrapper[4702]: I1125 14:02:21.970757 4702 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0f4314a6-9010-4789-816c-3bc52836c02e-util\") on node \"crc\" DevicePath \"\"" Nov 25 14:02:22 crc kubenswrapper[4702]: I1125 14:02:22.549421 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd" event={"ID":"0f4314a6-9010-4789-816c-3bc52836c02e","Type":"ContainerDied","Data":"1d2454cdd3804a8e2c52446c10de3cfe35c1bbb7dd75268e5c05a5291b2aea69"} Nov 25 14:02:22 crc kubenswrapper[4702]: I1125 14:02:22.549475 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d2454cdd3804a8e2c52446c10de3cfe35c1bbb7dd75268e5c05a5291b2aea69" Nov 25 14:02:22 crc kubenswrapper[4702]: I1125 14:02:22.549482 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd" Nov 25 14:02:31 crc kubenswrapper[4702]: I1125 14:02:31.610719 4702 generic.go:334] "Generic (PLEG): container finished" podID="3454fb60-87b9-45de-b8b2-13548d88cb1e" containerID="770cda36d12cd5d01cc54f5e98b263ef301e050bba16c09a4a3a2b4ca3be2a59" exitCode=0 Nov 25 14:02:31 crc kubenswrapper[4702]: I1125 14:02:31.610845 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"3454fb60-87b9-45de-b8b2-13548d88cb1e","Type":"ContainerDied","Data":"770cda36d12cd5d01cc54f5e98b263ef301e050bba16c09a4a3a2b4ca3be2a59"} Nov 25 14:02:32 crc kubenswrapper[4702]: I1125 14:02:32.085030 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-6b74b5c7-zfrjm"] Nov 25 14:02:32 crc kubenswrapper[4702]: E1125 14:02:32.085951 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f4314a6-9010-4789-816c-3bc52836c02e" containerName="pull" Nov 25 14:02:32 crc kubenswrapper[4702]: I1125 14:02:32.085975 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f4314a6-9010-4789-816c-3bc52836c02e" containerName="pull" Nov 25 14:02:32 crc kubenswrapper[4702]: E1125 14:02:32.086001 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f4314a6-9010-4789-816c-3bc52836c02e" containerName="extract" Nov 25 14:02:32 crc kubenswrapper[4702]: I1125 14:02:32.086009 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f4314a6-9010-4789-816c-3bc52836c02e" containerName="extract" Nov 25 14:02:32 crc kubenswrapper[4702]: E1125 14:02:32.086031 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f4314a6-9010-4789-816c-3bc52836c02e" containerName="util" Nov 25 14:02:32 crc kubenswrapper[4702]: I1125 14:02:32.086038 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f4314a6-9010-4789-816c-3bc52836c02e" containerName="util" Nov 25 14:02:32 crc kubenswrapper[4702]: I1125 14:02:32.086178 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f4314a6-9010-4789-816c-3bc52836c02e" containerName="extract" Nov 25 14:02:32 crc kubenswrapper[4702]: I1125 14:02:32.086808 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-6b74b5c7-zfrjm" Nov 25 14:02:32 crc kubenswrapper[4702]: I1125 14:02:32.093687 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-service-cert" Nov 25 14:02:32 crc kubenswrapper[4702]: I1125 14:02:32.093816 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-4tnsx" Nov 25 14:02:32 crc kubenswrapper[4702]: I1125 14:02:32.101728 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-6b74b5c7-zfrjm"] Nov 25 14:02:32 crc kubenswrapper[4702]: I1125 14:02:32.261205 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5b0aeb75-9843-4e68-9817-c1b0ac2062ea-apiservice-cert\") pod \"keystone-operator-controller-manager-6b74b5c7-zfrjm\" (UID: \"5b0aeb75-9843-4e68-9817-c1b0ac2062ea\") " pod="openstack-operators/keystone-operator-controller-manager-6b74b5c7-zfrjm" Nov 25 14:02:32 crc kubenswrapper[4702]: I1125 14:02:32.261300 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5b0aeb75-9843-4e68-9817-c1b0ac2062ea-webhook-cert\") pod \"keystone-operator-controller-manager-6b74b5c7-zfrjm\" (UID: \"5b0aeb75-9843-4e68-9817-c1b0ac2062ea\") " pod="openstack-operators/keystone-operator-controller-manager-6b74b5c7-zfrjm" Nov 25 14:02:32 crc kubenswrapper[4702]: I1125 14:02:32.261352 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h49wx\" (UniqueName: \"kubernetes.io/projected/5b0aeb75-9843-4e68-9817-c1b0ac2062ea-kube-api-access-h49wx\") pod \"keystone-operator-controller-manager-6b74b5c7-zfrjm\" (UID: \"5b0aeb75-9843-4e68-9817-c1b0ac2062ea\") " pod="openstack-operators/keystone-operator-controller-manager-6b74b5c7-zfrjm" Nov 25 14:02:32 crc kubenswrapper[4702]: I1125 14:02:32.364128 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h49wx\" (UniqueName: \"kubernetes.io/projected/5b0aeb75-9843-4e68-9817-c1b0ac2062ea-kube-api-access-h49wx\") pod \"keystone-operator-controller-manager-6b74b5c7-zfrjm\" (UID: \"5b0aeb75-9843-4e68-9817-c1b0ac2062ea\") " pod="openstack-operators/keystone-operator-controller-manager-6b74b5c7-zfrjm" Nov 25 14:02:32 crc kubenswrapper[4702]: I1125 14:02:32.364323 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5b0aeb75-9843-4e68-9817-c1b0ac2062ea-apiservice-cert\") pod \"keystone-operator-controller-manager-6b74b5c7-zfrjm\" (UID: \"5b0aeb75-9843-4e68-9817-c1b0ac2062ea\") " pod="openstack-operators/keystone-operator-controller-manager-6b74b5c7-zfrjm" Nov 25 14:02:32 crc kubenswrapper[4702]: I1125 14:02:32.364358 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5b0aeb75-9843-4e68-9817-c1b0ac2062ea-webhook-cert\") pod \"keystone-operator-controller-manager-6b74b5c7-zfrjm\" (UID: \"5b0aeb75-9843-4e68-9817-c1b0ac2062ea\") " pod="openstack-operators/keystone-operator-controller-manager-6b74b5c7-zfrjm" Nov 25 14:02:32 crc kubenswrapper[4702]: I1125 14:02:32.371011 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5b0aeb75-9843-4e68-9817-c1b0ac2062ea-webhook-cert\") pod \"keystone-operator-controller-manager-6b74b5c7-zfrjm\" (UID: \"5b0aeb75-9843-4e68-9817-c1b0ac2062ea\") " pod="openstack-operators/keystone-operator-controller-manager-6b74b5c7-zfrjm" Nov 25 14:02:32 crc kubenswrapper[4702]: I1125 14:02:32.383900 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5b0aeb75-9843-4e68-9817-c1b0ac2062ea-apiservice-cert\") pod \"keystone-operator-controller-manager-6b74b5c7-zfrjm\" (UID: \"5b0aeb75-9843-4e68-9817-c1b0ac2062ea\") " pod="openstack-operators/keystone-operator-controller-manager-6b74b5c7-zfrjm" Nov 25 14:02:32 crc kubenswrapper[4702]: I1125 14:02:32.387616 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h49wx\" (UniqueName: \"kubernetes.io/projected/5b0aeb75-9843-4e68-9817-c1b0ac2062ea-kube-api-access-h49wx\") pod \"keystone-operator-controller-manager-6b74b5c7-zfrjm\" (UID: \"5b0aeb75-9843-4e68-9817-c1b0ac2062ea\") " pod="openstack-operators/keystone-operator-controller-manager-6b74b5c7-zfrjm" Nov 25 14:02:32 crc kubenswrapper[4702]: I1125 14:02:32.405066 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-6b74b5c7-zfrjm" Nov 25 14:02:32 crc kubenswrapper[4702]: I1125 14:02:32.621019 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"3454fb60-87b9-45de-b8b2-13548d88cb1e","Type":"ContainerStarted","Data":"10ea28a9e4de6d53fa993cbcfb09c3f21596a01f1dbb100b2bb0d7d7646a9a4e"} Nov 25 14:02:32 crc kubenswrapper[4702]: I1125 14:02:32.621549 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 14:02:32 crc kubenswrapper[4702]: I1125 14:02:32.648916 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/rabbitmq-server-0" podStartSLOduration=37.148802817 podStartE2EDuration="45.64889934s" podCreationTimestamp="2025-11-25 14:01:47 +0000 UTC" firstStartedPulling="2025-11-25 14:01:49.366300974 +0000 UTC m=+918.528951424" lastFinishedPulling="2025-11-25 14:01:57.866397457 +0000 UTC m=+927.029047947" observedRunningTime="2025-11-25 14:02:32.640784496 +0000 UTC m=+961.803434966" watchObservedRunningTime="2025-11-25 14:02:32.64889934 +0000 UTC m=+961.811549790" Nov 25 14:02:32 crc kubenswrapper[4702]: I1125 14:02:32.833919 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-6b74b5c7-zfrjm"] Nov 25 14:02:32 crc kubenswrapper[4702]: W1125 14:02:32.856272 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b0aeb75_9843_4e68_9817_c1b0ac2062ea.slice/crio-4ad2885bf432194455ca7187c8ad9083caa8708c8542b62fbd63abe95dbece56 WatchSource:0}: Error finding container 4ad2885bf432194455ca7187c8ad9083caa8708c8542b62fbd63abe95dbece56: Status 404 returned error can't find the container with id 4ad2885bf432194455ca7187c8ad9083caa8708c8542b62fbd63abe95dbece56 Nov 25 14:02:32 crc kubenswrapper[4702]: I1125 14:02:32.859375 4702 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 14:02:33 crc kubenswrapper[4702]: I1125 14:02:33.631850 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-6b74b5c7-zfrjm" event={"ID":"5b0aeb75-9843-4e68-9817-c1b0ac2062ea","Type":"ContainerStarted","Data":"4ad2885bf432194455ca7187c8ad9083caa8708c8542b62fbd63abe95dbece56"} Nov 25 14:02:36 crc kubenswrapper[4702]: I1125 14:02:36.662610 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-6b74b5c7-zfrjm" event={"ID":"5b0aeb75-9843-4e68-9817-c1b0ac2062ea","Type":"ContainerStarted","Data":"e3a71ed8d602d1fc66849c2077494416462526afe6ee1606499524c4196b0a95"} Nov 25 14:02:36 crc kubenswrapper[4702]: I1125 14:02:36.663272 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-6b74b5c7-zfrjm" Nov 25 14:02:36 crc kubenswrapper[4702]: I1125 14:02:36.686710 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-6b74b5c7-zfrjm" podStartSLOduration=1.087125679 podStartE2EDuration="4.686686128s" podCreationTimestamp="2025-11-25 14:02:32 +0000 UTC" firstStartedPulling="2025-11-25 14:02:32.859021564 +0000 UTC m=+962.021672014" lastFinishedPulling="2025-11-25 14:02:36.458582013 +0000 UTC m=+965.621232463" observedRunningTime="2025-11-25 14:02:36.680590984 +0000 UTC m=+965.843241444" watchObservedRunningTime="2025-11-25 14:02:36.686686128 +0000 UTC m=+965.849336578" Nov 25 14:02:42 crc kubenswrapper[4702]: I1125 14:02:42.413969 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-6b74b5c7-zfrjm" Nov 25 14:02:43 crc kubenswrapper[4702]: I1125 14:02:43.256620 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-1d57-account-create-update-rnh45"] Nov 25 14:02:43 crc kubenswrapper[4702]: I1125 14:02:43.257880 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-1d57-account-create-update-rnh45" Nov 25 14:02:43 crc kubenswrapper[4702]: I1125 14:02:43.261054 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-create-smlsw"] Nov 25 14:02:43 crc kubenswrapper[4702]: I1125 14:02:43.261849 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-db-secret" Nov 25 14:02:43 crc kubenswrapper[4702]: I1125 14:02:43.261896 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-smlsw" Nov 25 14:02:43 crc kubenswrapper[4702]: I1125 14:02:43.267961 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-smlsw"] Nov 25 14:02:43 crc kubenswrapper[4702]: I1125 14:02:43.272736 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-1d57-account-create-update-rnh45"] Nov 25 14:02:43 crc kubenswrapper[4702]: I1125 14:02:43.429554 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzc7c\" (UniqueName: \"kubernetes.io/projected/9a144271-8855-4fba-8913-d9e1cef952d0-kube-api-access-bzc7c\") pod \"keystone-1d57-account-create-update-rnh45\" (UID: \"9a144271-8855-4fba-8913-d9e1cef952d0\") " pod="glance-kuttl-tests/keystone-1d57-account-create-update-rnh45" Nov 25 14:02:43 crc kubenswrapper[4702]: I1125 14:02:43.429608 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a144271-8855-4fba-8913-d9e1cef952d0-operator-scripts\") pod \"keystone-1d57-account-create-update-rnh45\" (UID: \"9a144271-8855-4fba-8913-d9e1cef952d0\") " pod="glance-kuttl-tests/keystone-1d57-account-create-update-rnh45" Nov 25 14:02:43 crc kubenswrapper[4702]: I1125 14:02:43.429700 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjpxs\" (UniqueName: \"kubernetes.io/projected/8d45b3f0-68f4-488f-b356-660f02a504ee-kube-api-access-xjpxs\") pod \"keystone-db-create-smlsw\" (UID: \"8d45b3f0-68f4-488f-b356-660f02a504ee\") " pod="glance-kuttl-tests/keystone-db-create-smlsw" Nov 25 14:02:43 crc kubenswrapper[4702]: I1125 14:02:43.429751 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d45b3f0-68f4-488f-b356-660f02a504ee-operator-scripts\") pod \"keystone-db-create-smlsw\" (UID: \"8d45b3f0-68f4-488f-b356-660f02a504ee\") " pod="glance-kuttl-tests/keystone-db-create-smlsw" Nov 25 14:02:43 crc kubenswrapper[4702]: I1125 14:02:43.530934 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzc7c\" (UniqueName: \"kubernetes.io/projected/9a144271-8855-4fba-8913-d9e1cef952d0-kube-api-access-bzc7c\") pod \"keystone-1d57-account-create-update-rnh45\" (UID: \"9a144271-8855-4fba-8913-d9e1cef952d0\") " pod="glance-kuttl-tests/keystone-1d57-account-create-update-rnh45" Nov 25 14:02:43 crc kubenswrapper[4702]: I1125 14:02:43.530994 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a144271-8855-4fba-8913-d9e1cef952d0-operator-scripts\") pod \"keystone-1d57-account-create-update-rnh45\" (UID: \"9a144271-8855-4fba-8913-d9e1cef952d0\") " pod="glance-kuttl-tests/keystone-1d57-account-create-update-rnh45" Nov 25 14:02:43 crc kubenswrapper[4702]: I1125 14:02:43.531052 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjpxs\" (UniqueName: \"kubernetes.io/projected/8d45b3f0-68f4-488f-b356-660f02a504ee-kube-api-access-xjpxs\") pod \"keystone-db-create-smlsw\" (UID: \"8d45b3f0-68f4-488f-b356-660f02a504ee\") " pod="glance-kuttl-tests/keystone-db-create-smlsw" Nov 25 14:02:43 crc kubenswrapper[4702]: I1125 14:02:43.531075 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d45b3f0-68f4-488f-b356-660f02a504ee-operator-scripts\") pod \"keystone-db-create-smlsw\" (UID: \"8d45b3f0-68f4-488f-b356-660f02a504ee\") " pod="glance-kuttl-tests/keystone-db-create-smlsw" Nov 25 14:02:43 crc kubenswrapper[4702]: I1125 14:02:43.532138 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a144271-8855-4fba-8913-d9e1cef952d0-operator-scripts\") pod \"keystone-1d57-account-create-update-rnh45\" (UID: \"9a144271-8855-4fba-8913-d9e1cef952d0\") " pod="glance-kuttl-tests/keystone-1d57-account-create-update-rnh45" Nov 25 14:02:43 crc kubenswrapper[4702]: I1125 14:02:43.532157 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d45b3f0-68f4-488f-b356-660f02a504ee-operator-scripts\") pod \"keystone-db-create-smlsw\" (UID: \"8d45b3f0-68f4-488f-b356-660f02a504ee\") " pod="glance-kuttl-tests/keystone-db-create-smlsw" Nov 25 14:02:43 crc kubenswrapper[4702]: I1125 14:02:43.555747 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzc7c\" (UniqueName: \"kubernetes.io/projected/9a144271-8855-4fba-8913-d9e1cef952d0-kube-api-access-bzc7c\") pod \"keystone-1d57-account-create-update-rnh45\" (UID: \"9a144271-8855-4fba-8913-d9e1cef952d0\") " pod="glance-kuttl-tests/keystone-1d57-account-create-update-rnh45" Nov 25 14:02:43 crc kubenswrapper[4702]: I1125 14:02:43.564300 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjpxs\" (UniqueName: \"kubernetes.io/projected/8d45b3f0-68f4-488f-b356-660f02a504ee-kube-api-access-xjpxs\") pod \"keystone-db-create-smlsw\" (UID: \"8d45b3f0-68f4-488f-b356-660f02a504ee\") " pod="glance-kuttl-tests/keystone-db-create-smlsw" Nov 25 14:02:43 crc kubenswrapper[4702]: I1125 14:02:43.576847 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-1d57-account-create-update-rnh45" Nov 25 14:02:43 crc kubenswrapper[4702]: I1125 14:02:43.585491 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-smlsw" Nov 25 14:02:44 crc kubenswrapper[4702]: I1125 14:02:44.093081 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-smlsw"] Nov 25 14:02:44 crc kubenswrapper[4702]: I1125 14:02:44.211891 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-1d57-account-create-update-rnh45"] Nov 25 14:02:44 crc kubenswrapper[4702]: I1125 14:02:44.719000 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-1d57-account-create-update-rnh45" event={"ID":"9a144271-8855-4fba-8913-d9e1cef952d0","Type":"ContainerStarted","Data":"022275568c8772696cf7fba610a0bcdb3ce8f631d918ee164271905aa1bf0b10"} Nov 25 14:02:44 crc kubenswrapper[4702]: I1125 14:02:44.719427 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-1d57-account-create-update-rnh45" event={"ID":"9a144271-8855-4fba-8913-d9e1cef952d0","Type":"ContainerStarted","Data":"8b75008877bcbe4e23858cb16b8e4a24ba2eba68e1fc19a834af3a873d23a556"} Nov 25 14:02:44 crc kubenswrapper[4702]: I1125 14:02:44.721039 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-smlsw" event={"ID":"8d45b3f0-68f4-488f-b356-660f02a504ee","Type":"ContainerStarted","Data":"c693c08b947e18d663547be50ad70954bc994cf9faf3218e24e1713f65ea23a2"} Nov 25 14:02:44 crc kubenswrapper[4702]: I1125 14:02:44.721088 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-smlsw" event={"ID":"8d45b3f0-68f4-488f-b356-660f02a504ee","Type":"ContainerStarted","Data":"e7975b34d7a1a64403f05dd864526264cc721658e105b0aee9359c64f24aa605"} Nov 25 14:02:44 crc kubenswrapper[4702]: I1125 14:02:44.743256 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-1d57-account-create-update-rnh45" podStartSLOduration=1.743220666 podStartE2EDuration="1.743220666s" podCreationTimestamp="2025-11-25 14:02:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 14:02:44.738785134 +0000 UTC m=+973.901435584" watchObservedRunningTime="2025-11-25 14:02:44.743220666 +0000 UTC m=+973.905871116" Nov 25 14:02:45 crc kubenswrapper[4702]: I1125 14:02:45.730993 4702 generic.go:334] "Generic (PLEG): container finished" podID="8d45b3f0-68f4-488f-b356-660f02a504ee" containerID="c693c08b947e18d663547be50ad70954bc994cf9faf3218e24e1713f65ea23a2" exitCode=0 Nov 25 14:02:45 crc kubenswrapper[4702]: I1125 14:02:45.731064 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-smlsw" event={"ID":"8d45b3f0-68f4-488f-b356-660f02a504ee","Type":"ContainerDied","Data":"c693c08b947e18d663547be50ad70954bc994cf9faf3218e24e1713f65ea23a2"} Nov 25 14:02:45 crc kubenswrapper[4702]: I1125 14:02:45.992704 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-index-l2sm2"] Nov 25 14:02:45 crc kubenswrapper[4702]: I1125 14:02:45.994159 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-l2sm2" Nov 25 14:02:45 crc kubenswrapper[4702]: I1125 14:02:45.996941 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-index-dockercfg-grpbf" Nov 25 14:02:45 crc kubenswrapper[4702]: I1125 14:02:45.999033 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-l2sm2"] Nov 25 14:02:46 crc kubenswrapper[4702]: I1125 14:02:46.181069 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2pck\" (UniqueName: \"kubernetes.io/projected/62409d84-e1be-4265-ae7b-b58face7988b-kube-api-access-f2pck\") pod \"horizon-operator-index-l2sm2\" (UID: \"62409d84-e1be-4265-ae7b-b58face7988b\") " pod="openstack-operators/horizon-operator-index-l2sm2" Nov 25 14:02:46 crc kubenswrapper[4702]: I1125 14:02:46.282449 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2pck\" (UniqueName: \"kubernetes.io/projected/62409d84-e1be-4265-ae7b-b58face7988b-kube-api-access-f2pck\") pod \"horizon-operator-index-l2sm2\" (UID: \"62409d84-e1be-4265-ae7b-b58face7988b\") " pod="openstack-operators/horizon-operator-index-l2sm2" Nov 25 14:02:46 crc kubenswrapper[4702]: I1125 14:02:46.306476 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2pck\" (UniqueName: \"kubernetes.io/projected/62409d84-e1be-4265-ae7b-b58face7988b-kube-api-access-f2pck\") pod \"horizon-operator-index-l2sm2\" (UID: \"62409d84-e1be-4265-ae7b-b58face7988b\") " pod="openstack-operators/horizon-operator-index-l2sm2" Nov 25 14:02:46 crc kubenswrapper[4702]: I1125 14:02:46.315285 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-l2sm2" Nov 25 14:02:46 crc kubenswrapper[4702]: I1125 14:02:46.744246 4702 generic.go:334] "Generic (PLEG): container finished" podID="9a144271-8855-4fba-8913-d9e1cef952d0" containerID="022275568c8772696cf7fba610a0bcdb3ce8f631d918ee164271905aa1bf0b10" exitCode=0 Nov 25 14:02:46 crc kubenswrapper[4702]: I1125 14:02:46.744277 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-1d57-account-create-update-rnh45" event={"ID":"9a144271-8855-4fba-8913-d9e1cef952d0","Type":"ContainerDied","Data":"022275568c8772696cf7fba610a0bcdb3ce8f631d918ee164271905aa1bf0b10"} Nov 25 14:02:46 crc kubenswrapper[4702]: I1125 14:02:46.760257 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-l2sm2"] Nov 25 14:02:46 crc kubenswrapper[4702]: W1125 14:02:46.776375 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62409d84_e1be_4265_ae7b_b58face7988b.slice/crio-a7ef15bf601e8ee05af87993f514af08925850c9fc22663f8bd6d97c8fab38a1 WatchSource:0}: Error finding container a7ef15bf601e8ee05af87993f514af08925850c9fc22663f8bd6d97c8fab38a1: Status 404 returned error can't find the container with id a7ef15bf601e8ee05af87993f514af08925850c9fc22663f8bd6d97c8fab38a1 Nov 25 14:02:47 crc kubenswrapper[4702]: I1125 14:02:47.028309 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-smlsw" Nov 25 14:02:47 crc kubenswrapper[4702]: I1125 14:02:47.196073 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjpxs\" (UniqueName: \"kubernetes.io/projected/8d45b3f0-68f4-488f-b356-660f02a504ee-kube-api-access-xjpxs\") pod \"8d45b3f0-68f4-488f-b356-660f02a504ee\" (UID: \"8d45b3f0-68f4-488f-b356-660f02a504ee\") " Nov 25 14:02:47 crc kubenswrapper[4702]: I1125 14:02:47.196263 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d45b3f0-68f4-488f-b356-660f02a504ee-operator-scripts\") pod \"8d45b3f0-68f4-488f-b356-660f02a504ee\" (UID: \"8d45b3f0-68f4-488f-b356-660f02a504ee\") " Nov 25 14:02:47 crc kubenswrapper[4702]: I1125 14:02:47.196940 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d45b3f0-68f4-488f-b356-660f02a504ee-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8d45b3f0-68f4-488f-b356-660f02a504ee" (UID: "8d45b3f0-68f4-488f-b356-660f02a504ee"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 14:02:47 crc kubenswrapper[4702]: I1125 14:02:47.204412 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d45b3f0-68f4-488f-b356-660f02a504ee-kube-api-access-xjpxs" (OuterVolumeSpecName: "kube-api-access-xjpxs") pod "8d45b3f0-68f4-488f-b356-660f02a504ee" (UID: "8d45b3f0-68f4-488f-b356-660f02a504ee"). InnerVolumeSpecName "kube-api-access-xjpxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:02:47 crc kubenswrapper[4702]: I1125 14:02:47.297840 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjpxs\" (UniqueName: \"kubernetes.io/projected/8d45b3f0-68f4-488f-b356-660f02a504ee-kube-api-access-xjpxs\") on node \"crc\" DevicePath \"\"" Nov 25 14:02:47 crc kubenswrapper[4702]: I1125 14:02:47.297882 4702 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d45b3f0-68f4-488f-b356-660f02a504ee-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 14:02:47 crc kubenswrapper[4702]: I1125 14:02:47.754433 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-smlsw" event={"ID":"8d45b3f0-68f4-488f-b356-660f02a504ee","Type":"ContainerDied","Data":"e7975b34d7a1a64403f05dd864526264cc721658e105b0aee9359c64f24aa605"} Nov 25 14:02:47 crc kubenswrapper[4702]: I1125 14:02:47.754485 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e7975b34d7a1a64403f05dd864526264cc721658e105b0aee9359c64f24aa605" Nov 25 14:02:47 crc kubenswrapper[4702]: I1125 14:02:47.754454 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-smlsw" Nov 25 14:02:47 crc kubenswrapper[4702]: I1125 14:02:47.756342 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-l2sm2" event={"ID":"62409d84-e1be-4265-ae7b-b58face7988b","Type":"ContainerStarted","Data":"a7ef15bf601e8ee05af87993f514af08925850c9fc22663f8bd6d97c8fab38a1"} Nov 25 14:02:48 crc kubenswrapper[4702]: I1125 14:02:48.021663 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-1d57-account-create-update-rnh45" Nov 25 14:02:48 crc kubenswrapper[4702]: I1125 14:02:48.111023 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a144271-8855-4fba-8913-d9e1cef952d0-operator-scripts\") pod \"9a144271-8855-4fba-8913-d9e1cef952d0\" (UID: \"9a144271-8855-4fba-8913-d9e1cef952d0\") " Nov 25 14:02:48 crc kubenswrapper[4702]: I1125 14:02:48.111086 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzc7c\" (UniqueName: \"kubernetes.io/projected/9a144271-8855-4fba-8913-d9e1cef952d0-kube-api-access-bzc7c\") pod \"9a144271-8855-4fba-8913-d9e1cef952d0\" (UID: \"9a144271-8855-4fba-8913-d9e1cef952d0\") " Nov 25 14:02:48 crc kubenswrapper[4702]: I1125 14:02:48.113164 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a144271-8855-4fba-8913-d9e1cef952d0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9a144271-8855-4fba-8913-d9e1cef952d0" (UID: "9a144271-8855-4fba-8913-d9e1cef952d0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 14:02:48 crc kubenswrapper[4702]: I1125 14:02:48.126326 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a144271-8855-4fba-8913-d9e1cef952d0-kube-api-access-bzc7c" (OuterVolumeSpecName: "kube-api-access-bzc7c") pod "9a144271-8855-4fba-8913-d9e1cef952d0" (UID: "9a144271-8855-4fba-8913-d9e1cef952d0"). InnerVolumeSpecName "kube-api-access-bzc7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:02:48 crc kubenswrapper[4702]: I1125 14:02:48.216438 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzc7c\" (UniqueName: \"kubernetes.io/projected/9a144271-8855-4fba-8913-d9e1cef952d0-kube-api-access-bzc7c\") on node \"crc\" DevicePath \"\"" Nov 25 14:02:48 crc kubenswrapper[4702]: I1125 14:02:48.216475 4702 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a144271-8855-4fba-8913-d9e1cef952d0-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 14:02:48 crc kubenswrapper[4702]: I1125 14:02:48.764364 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-1d57-account-create-update-rnh45" event={"ID":"9a144271-8855-4fba-8913-d9e1cef952d0","Type":"ContainerDied","Data":"8b75008877bcbe4e23858cb16b8e4a24ba2eba68e1fc19a834af3a873d23a556"} Nov 25 14:02:48 crc kubenswrapper[4702]: I1125 14:02:48.764415 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b75008877bcbe4e23858cb16b8e4a24ba2eba68e1fc19a834af3a873d23a556" Nov 25 14:02:48 crc kubenswrapper[4702]: I1125 14:02:48.764472 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-1d57-account-create-update-rnh45" Nov 25 14:02:49 crc kubenswrapper[4702]: I1125 14:02:49.100415 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 14:02:49 crc kubenswrapper[4702]: I1125 14:02:49.587684 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-index-9htnh"] Nov 25 14:02:49 crc kubenswrapper[4702]: E1125 14:02:49.588259 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a144271-8855-4fba-8913-d9e1cef952d0" containerName="mariadb-account-create-update" Nov 25 14:02:49 crc kubenswrapper[4702]: I1125 14:02:49.588276 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a144271-8855-4fba-8913-d9e1cef952d0" containerName="mariadb-account-create-update" Nov 25 14:02:49 crc kubenswrapper[4702]: E1125 14:02:49.588306 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d45b3f0-68f4-488f-b356-660f02a504ee" containerName="mariadb-database-create" Nov 25 14:02:49 crc kubenswrapper[4702]: I1125 14:02:49.588313 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d45b3f0-68f4-488f-b356-660f02a504ee" containerName="mariadb-database-create" Nov 25 14:02:49 crc kubenswrapper[4702]: I1125 14:02:49.588421 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a144271-8855-4fba-8913-d9e1cef952d0" containerName="mariadb-account-create-update" Nov 25 14:02:49 crc kubenswrapper[4702]: I1125 14:02:49.588436 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d45b3f0-68f4-488f-b356-660f02a504ee" containerName="mariadb-database-create" Nov 25 14:02:49 crc kubenswrapper[4702]: I1125 14:02:49.588846 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-9htnh" Nov 25 14:02:49 crc kubenswrapper[4702]: I1125 14:02:49.593091 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-index-dockercfg-n7ctg" Nov 25 14:02:49 crc kubenswrapper[4702]: I1125 14:02:49.603212 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-9htnh"] Nov 25 14:02:49 crc kubenswrapper[4702]: I1125 14:02:49.741370 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69ccb\" (UniqueName: \"kubernetes.io/projected/61fcd26c-69b0-4a87-951e-7a4f002acff9-kube-api-access-69ccb\") pod \"swift-operator-index-9htnh\" (UID: \"61fcd26c-69b0-4a87-951e-7a4f002acff9\") " pod="openstack-operators/swift-operator-index-9htnh" Nov 25 14:02:49 crc kubenswrapper[4702]: I1125 14:02:49.842859 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69ccb\" (UniqueName: \"kubernetes.io/projected/61fcd26c-69b0-4a87-951e-7a4f002acff9-kube-api-access-69ccb\") pod \"swift-operator-index-9htnh\" (UID: \"61fcd26c-69b0-4a87-951e-7a4f002acff9\") " pod="openstack-operators/swift-operator-index-9htnh" Nov 25 14:02:49 crc kubenswrapper[4702]: I1125 14:02:49.864223 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69ccb\" (UniqueName: \"kubernetes.io/projected/61fcd26c-69b0-4a87-951e-7a4f002acff9-kube-api-access-69ccb\") pod \"swift-operator-index-9htnh\" (UID: \"61fcd26c-69b0-4a87-951e-7a4f002acff9\") " pod="openstack-operators/swift-operator-index-9htnh" Nov 25 14:02:49 crc kubenswrapper[4702]: I1125 14:02:49.905051 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-9htnh" Nov 25 14:02:50 crc kubenswrapper[4702]: I1125 14:02:50.681997 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-9htnh"] Nov 25 14:02:50 crc kubenswrapper[4702]: W1125 14:02:50.691470 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod61fcd26c_69b0_4a87_951e_7a4f002acff9.slice/crio-3d5a3d138484d28e721a11b252b4066055862ce7195fcdb25ce3c520f55bb9e5 WatchSource:0}: Error finding container 3d5a3d138484d28e721a11b252b4066055862ce7195fcdb25ce3c520f55bb9e5: Status 404 returned error can't find the container with id 3d5a3d138484d28e721a11b252b4066055862ce7195fcdb25ce3c520f55bb9e5 Nov 25 14:02:50 crc kubenswrapper[4702]: I1125 14:02:50.780313 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-l2sm2" event={"ID":"62409d84-e1be-4265-ae7b-b58face7988b","Type":"ContainerStarted","Data":"c7022ef8841d6e6935e2246d8cd8203c6dc589862c1565140fedb3030d3c92ad"} Nov 25 14:02:50 crc kubenswrapper[4702]: I1125 14:02:50.781944 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-9htnh" event={"ID":"61fcd26c-69b0-4a87-951e-7a4f002acff9","Type":"ContainerStarted","Data":"3d5a3d138484d28e721a11b252b4066055862ce7195fcdb25ce3c520f55bb9e5"} Nov 25 14:02:50 crc kubenswrapper[4702]: I1125 14:02:50.802203 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-index-l2sm2" podStartSLOduration=2.321771298 podStartE2EDuration="5.802181331s" podCreationTimestamp="2025-11-25 14:02:45 +0000 UTC" firstStartedPulling="2025-11-25 14:02:46.786517987 +0000 UTC m=+975.949168457" lastFinishedPulling="2025-11-25 14:02:50.26692804 +0000 UTC m=+979.429578490" observedRunningTime="2025-11-25 14:02:50.795860843 +0000 UTC m=+979.958511293" watchObservedRunningTime="2025-11-25 14:02:50.802181331 +0000 UTC m=+979.964831801" Nov 25 14:02:51 crc kubenswrapper[4702]: I1125 14:02:51.789358 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-9htnh" event={"ID":"61fcd26c-69b0-4a87-951e-7a4f002acff9","Type":"ContainerStarted","Data":"1bee495db8af316716c6fce1fa8058c575a34aede527dab994e42071d907a599"} Nov 25 14:02:51 crc kubenswrapper[4702]: I1125 14:02:51.808327 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-index-9htnh" podStartSLOduration=1.973872909 podStartE2EDuration="2.808269657s" podCreationTimestamp="2025-11-25 14:02:49 +0000 UTC" firstStartedPulling="2025-11-25 14:02:50.694481008 +0000 UTC m=+979.857131458" lastFinishedPulling="2025-11-25 14:02:51.528877756 +0000 UTC m=+980.691528206" observedRunningTime="2025-11-25 14:02:51.807640081 +0000 UTC m=+980.970290531" watchObservedRunningTime="2025-11-25 14:02:51.808269657 +0000 UTC m=+980.970920107" Nov 25 14:02:53 crc kubenswrapper[4702]: I1125 14:02:53.824908 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-sync-5j9zq"] Nov 25 14:02:53 crc kubenswrapper[4702]: I1125 14:02:53.826459 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-5j9zq" Nov 25 14:02:53 crc kubenswrapper[4702]: I1125 14:02:53.829061 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 25 14:02:53 crc kubenswrapper[4702]: I1125 14:02:53.829064 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-42ddw" Nov 25 14:02:53 crc kubenswrapper[4702]: I1125 14:02:53.832688 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 25 14:02:53 crc kubenswrapper[4702]: I1125 14:02:53.832895 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 25 14:02:53 crc kubenswrapper[4702]: I1125 14:02:53.841666 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-5j9zq"] Nov 25 14:02:53 crc kubenswrapper[4702]: I1125 14:02:53.931421 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slxsb\" (UniqueName: \"kubernetes.io/projected/3339fdec-c17b-4052-a472-7c6e19475fe4-kube-api-access-slxsb\") pod \"keystone-db-sync-5j9zq\" (UID: \"3339fdec-c17b-4052-a472-7c6e19475fe4\") " pod="glance-kuttl-tests/keystone-db-sync-5j9zq" Nov 25 14:02:53 crc kubenswrapper[4702]: I1125 14:02:53.931528 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3339fdec-c17b-4052-a472-7c6e19475fe4-config-data\") pod \"keystone-db-sync-5j9zq\" (UID: \"3339fdec-c17b-4052-a472-7c6e19475fe4\") " pod="glance-kuttl-tests/keystone-db-sync-5j9zq" Nov 25 14:02:54 crc kubenswrapper[4702]: I1125 14:02:54.033010 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slxsb\" (UniqueName: \"kubernetes.io/projected/3339fdec-c17b-4052-a472-7c6e19475fe4-kube-api-access-slxsb\") pod \"keystone-db-sync-5j9zq\" (UID: \"3339fdec-c17b-4052-a472-7c6e19475fe4\") " pod="glance-kuttl-tests/keystone-db-sync-5j9zq" Nov 25 14:02:54 crc kubenswrapper[4702]: I1125 14:02:54.033120 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3339fdec-c17b-4052-a472-7c6e19475fe4-config-data\") pod \"keystone-db-sync-5j9zq\" (UID: \"3339fdec-c17b-4052-a472-7c6e19475fe4\") " pod="glance-kuttl-tests/keystone-db-sync-5j9zq" Nov 25 14:02:54 crc kubenswrapper[4702]: I1125 14:02:54.044052 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3339fdec-c17b-4052-a472-7c6e19475fe4-config-data\") pod \"keystone-db-sync-5j9zq\" (UID: \"3339fdec-c17b-4052-a472-7c6e19475fe4\") " pod="glance-kuttl-tests/keystone-db-sync-5j9zq" Nov 25 14:02:54 crc kubenswrapper[4702]: I1125 14:02:54.053948 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slxsb\" (UniqueName: \"kubernetes.io/projected/3339fdec-c17b-4052-a472-7c6e19475fe4-kube-api-access-slxsb\") pod \"keystone-db-sync-5j9zq\" (UID: \"3339fdec-c17b-4052-a472-7c6e19475fe4\") " pod="glance-kuttl-tests/keystone-db-sync-5j9zq" Nov 25 14:02:54 crc kubenswrapper[4702]: I1125 14:02:54.148610 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-5j9zq" Nov 25 14:02:54 crc kubenswrapper[4702]: I1125 14:02:54.609904 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-5j9zq"] Nov 25 14:02:54 crc kubenswrapper[4702]: W1125 14:02:54.621999 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3339fdec_c17b_4052_a472_7c6e19475fe4.slice/crio-f579e6a25e1a47ea270d84b9cb1bdce50ec12b8bbd3991528f7657aa4d592f7e WatchSource:0}: Error finding container f579e6a25e1a47ea270d84b9cb1bdce50ec12b8bbd3991528f7657aa4d592f7e: Status 404 returned error can't find the container with id f579e6a25e1a47ea270d84b9cb1bdce50ec12b8bbd3991528f7657aa4d592f7e Nov 25 14:02:54 crc kubenswrapper[4702]: I1125 14:02:54.817371 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-5j9zq" event={"ID":"3339fdec-c17b-4052-a472-7c6e19475fe4","Type":"ContainerStarted","Data":"f579e6a25e1a47ea270d84b9cb1bdce50ec12b8bbd3991528f7657aa4d592f7e"} Nov 25 14:02:56 crc kubenswrapper[4702]: I1125 14:02:56.315797 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/horizon-operator-index-l2sm2" Nov 25 14:02:56 crc kubenswrapper[4702]: I1125 14:02:56.316283 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-index-l2sm2" Nov 25 14:02:56 crc kubenswrapper[4702]: I1125 14:02:56.347318 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/horizon-operator-index-l2sm2" Nov 25 14:02:56 crc kubenswrapper[4702]: I1125 14:02:56.871557 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-index-l2sm2" Nov 25 14:02:59 crc kubenswrapper[4702]: I1125 14:02:59.906210 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-index-9htnh" Nov 25 14:02:59 crc kubenswrapper[4702]: I1125 14:02:59.907275 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/swift-operator-index-9htnh" Nov 25 14:02:59 crc kubenswrapper[4702]: I1125 14:02:59.939965 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/swift-operator-index-9htnh" Nov 25 14:03:00 crc kubenswrapper[4702]: I1125 14:03:00.896595 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-index-9htnh" Nov 25 14:03:02 crc kubenswrapper[4702]: I1125 14:03:02.883132 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-5j9zq" event={"ID":"3339fdec-c17b-4052-a472-7c6e19475fe4","Type":"ContainerStarted","Data":"ab70360378ad003c37db0e35718f2a0c9d8610dab59da2005e4185453011362e"} Nov 25 14:03:02 crc kubenswrapper[4702]: I1125 14:03:02.905091 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-db-sync-5j9zq" podStartSLOduration=2.5412022739999998 podStartE2EDuration="9.905063534s" podCreationTimestamp="2025-11-25 14:02:53 +0000 UTC" firstStartedPulling="2025-11-25 14:02:54.625081517 +0000 UTC m=+983.787731967" lastFinishedPulling="2025-11-25 14:03:01.988942777 +0000 UTC m=+991.151593227" observedRunningTime="2025-11-25 14:03:02.900600103 +0000 UTC m=+992.063250563" watchObservedRunningTime="2025-11-25 14:03:02.905063534 +0000 UTC m=+992.067714004" Nov 25 14:03:03 crc kubenswrapper[4702]: I1125 14:03:03.962947 4702 patch_prober.go:28] interesting pod/machine-config-daemon-qqr76 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 14:03:03 crc kubenswrapper[4702]: I1125 14:03:03.963384 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 14:03:06 crc kubenswrapper[4702]: I1125 14:03:06.452408 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp"] Nov 25 14:03:06 crc kubenswrapper[4702]: I1125 14:03:06.454043 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp" Nov 25 14:03:06 crc kubenswrapper[4702]: I1125 14:03:06.456528 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-6r4jq" Nov 25 14:03:06 crc kubenswrapper[4702]: I1125 14:03:06.469318 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp"] Nov 25 14:03:06 crc kubenswrapper[4702]: I1125 14:03:06.642714 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/015eb990-26fa-467f-b645-fd3290426356-bundle\") pod \"aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp\" (UID: \"015eb990-26fa-467f-b645-fd3290426356\") " pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp" Nov 25 14:03:06 crc kubenswrapper[4702]: I1125 14:03:06.643137 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/015eb990-26fa-467f-b645-fd3290426356-util\") pod \"aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp\" (UID: \"015eb990-26fa-467f-b645-fd3290426356\") " pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp" Nov 25 14:03:06 crc kubenswrapper[4702]: I1125 14:03:06.643260 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngxqg\" (UniqueName: \"kubernetes.io/projected/015eb990-26fa-467f-b645-fd3290426356-kube-api-access-ngxqg\") pod \"aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp\" (UID: \"015eb990-26fa-467f-b645-fd3290426356\") " pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp" Nov 25 14:03:06 crc kubenswrapper[4702]: I1125 14:03:06.745113 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/015eb990-26fa-467f-b645-fd3290426356-bundle\") pod \"aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp\" (UID: \"015eb990-26fa-467f-b645-fd3290426356\") " pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp" Nov 25 14:03:06 crc kubenswrapper[4702]: I1125 14:03:06.745200 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/015eb990-26fa-467f-b645-fd3290426356-util\") pod \"aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp\" (UID: \"015eb990-26fa-467f-b645-fd3290426356\") " pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp" Nov 25 14:03:06 crc kubenswrapper[4702]: I1125 14:03:06.745250 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngxqg\" (UniqueName: \"kubernetes.io/projected/015eb990-26fa-467f-b645-fd3290426356-kube-api-access-ngxqg\") pod \"aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp\" (UID: \"015eb990-26fa-467f-b645-fd3290426356\") " pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp" Nov 25 14:03:06 crc kubenswrapper[4702]: I1125 14:03:06.746390 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/015eb990-26fa-467f-b645-fd3290426356-bundle\") pod \"aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp\" (UID: \"015eb990-26fa-467f-b645-fd3290426356\") " pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp" Nov 25 14:03:06 crc kubenswrapper[4702]: I1125 14:03:06.746674 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/015eb990-26fa-467f-b645-fd3290426356-util\") pod \"aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp\" (UID: \"015eb990-26fa-467f-b645-fd3290426356\") " pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp" Nov 25 14:03:06 crc kubenswrapper[4702]: I1125 14:03:06.768210 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngxqg\" (UniqueName: \"kubernetes.io/projected/015eb990-26fa-467f-b645-fd3290426356-kube-api-access-ngxqg\") pod \"aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp\" (UID: \"015eb990-26fa-467f-b645-fd3290426356\") " pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp" Nov 25 14:03:06 crc kubenswrapper[4702]: I1125 14:03:06.778221 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp" Nov 25 14:03:07 crc kubenswrapper[4702]: I1125 14:03:07.244389 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp"] Nov 25 14:03:07 crc kubenswrapper[4702]: I1125 14:03:07.250156 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9"] Nov 25 14:03:07 crc kubenswrapper[4702]: I1125 14:03:07.252889 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9" Nov 25 14:03:07 crc kubenswrapper[4702]: I1125 14:03:07.261256 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9"] Nov 25 14:03:07 crc kubenswrapper[4702]: I1125 14:03:07.353796 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbd68\" (UniqueName: \"kubernetes.io/projected/401b6536-9504-448e-8acb-cb09c0cb0650-kube-api-access-bbd68\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9\" (UID: \"401b6536-9504-448e-8acb-cb09c0cb0650\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9" Nov 25 14:03:07 crc kubenswrapper[4702]: I1125 14:03:07.353867 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/401b6536-9504-448e-8acb-cb09c0cb0650-bundle\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9\" (UID: \"401b6536-9504-448e-8acb-cb09c0cb0650\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9" Nov 25 14:03:07 crc kubenswrapper[4702]: I1125 14:03:07.353941 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/401b6536-9504-448e-8acb-cb09c0cb0650-util\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9\" (UID: \"401b6536-9504-448e-8acb-cb09c0cb0650\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9" Nov 25 14:03:07 crc kubenswrapper[4702]: I1125 14:03:07.456361 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/401b6536-9504-448e-8acb-cb09c0cb0650-bundle\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9\" (UID: \"401b6536-9504-448e-8acb-cb09c0cb0650\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9" Nov 25 14:03:07 crc kubenswrapper[4702]: I1125 14:03:07.456497 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/401b6536-9504-448e-8acb-cb09c0cb0650-util\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9\" (UID: \"401b6536-9504-448e-8acb-cb09c0cb0650\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9" Nov 25 14:03:07 crc kubenswrapper[4702]: I1125 14:03:07.456595 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbd68\" (UniqueName: \"kubernetes.io/projected/401b6536-9504-448e-8acb-cb09c0cb0650-kube-api-access-bbd68\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9\" (UID: \"401b6536-9504-448e-8acb-cb09c0cb0650\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9" Nov 25 14:03:07 crc kubenswrapper[4702]: I1125 14:03:07.457092 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/401b6536-9504-448e-8acb-cb09c0cb0650-util\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9\" (UID: \"401b6536-9504-448e-8acb-cb09c0cb0650\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9" Nov 25 14:03:07 crc kubenswrapper[4702]: I1125 14:03:07.457160 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/401b6536-9504-448e-8acb-cb09c0cb0650-bundle\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9\" (UID: \"401b6536-9504-448e-8acb-cb09c0cb0650\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9" Nov 25 14:03:07 crc kubenswrapper[4702]: I1125 14:03:07.480119 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbd68\" (UniqueName: \"kubernetes.io/projected/401b6536-9504-448e-8acb-cb09c0cb0650-kube-api-access-bbd68\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9\" (UID: \"401b6536-9504-448e-8acb-cb09c0cb0650\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9" Nov 25 14:03:07 crc kubenswrapper[4702]: I1125 14:03:07.720549 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9" Nov 25 14:03:07 crc kubenswrapper[4702]: I1125 14:03:07.926336 4702 generic.go:334] "Generic (PLEG): container finished" podID="015eb990-26fa-467f-b645-fd3290426356" containerID="513cbe6af894cc36814c46731c1dedeb57d067f897132cfa9be536e8f2e6ab82" exitCode=0 Nov 25 14:03:07 crc kubenswrapper[4702]: I1125 14:03:07.926391 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp" event={"ID":"015eb990-26fa-467f-b645-fd3290426356","Type":"ContainerDied","Data":"513cbe6af894cc36814c46731c1dedeb57d067f897132cfa9be536e8f2e6ab82"} Nov 25 14:03:07 crc kubenswrapper[4702]: I1125 14:03:07.926449 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp" event={"ID":"015eb990-26fa-467f-b645-fd3290426356","Type":"ContainerStarted","Data":"8ff7a2bb3071b6b5ccaa589e5a3b17edf35c464c377d77d06ed0ec03d631be07"} Nov 25 14:03:08 crc kubenswrapper[4702]: I1125 14:03:08.142430 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9"] Nov 25 14:03:08 crc kubenswrapper[4702]: I1125 14:03:08.936468 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9" event={"ID":"401b6536-9504-448e-8acb-cb09c0cb0650","Type":"ContainerStarted","Data":"fb151adb73a593fcee10fc70d7a09ca15ce479032ef169b502b18840e521cad0"} Nov 25 14:03:10 crc kubenswrapper[4702]: I1125 14:03:10.959197 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9" event={"ID":"401b6536-9504-448e-8acb-cb09c0cb0650","Type":"ContainerStarted","Data":"2568ce2377d0e0d40052aacfa2567ae07327967881ab74d0d9f1d120d38d6977"} Nov 25 14:03:11 crc kubenswrapper[4702]: I1125 14:03:11.966536 4702 generic.go:334] "Generic (PLEG): container finished" podID="401b6536-9504-448e-8acb-cb09c0cb0650" containerID="2568ce2377d0e0d40052aacfa2567ae07327967881ab74d0d9f1d120d38d6977" exitCode=0 Nov 25 14:03:11 crc kubenswrapper[4702]: I1125 14:03:11.966656 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9" event={"ID":"401b6536-9504-448e-8acb-cb09c0cb0650","Type":"ContainerDied","Data":"2568ce2377d0e0d40052aacfa2567ae07327967881ab74d0d9f1d120d38d6977"} Nov 25 14:03:15 crc kubenswrapper[4702]: I1125 14:03:15.995843 4702 generic.go:334] "Generic (PLEG): container finished" podID="015eb990-26fa-467f-b645-fd3290426356" containerID="0826bfdb16e62704b278fbf97143d82e59d07e9c36f55342408a59c6158c01ea" exitCode=0 Nov 25 14:03:15 crc kubenswrapper[4702]: I1125 14:03:15.995897 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp" event={"ID":"015eb990-26fa-467f-b645-fd3290426356","Type":"ContainerDied","Data":"0826bfdb16e62704b278fbf97143d82e59d07e9c36f55342408a59c6158c01ea"} Nov 25 14:03:17 crc kubenswrapper[4702]: I1125 14:03:17.008941 4702 generic.go:334] "Generic (PLEG): container finished" podID="401b6536-9504-448e-8acb-cb09c0cb0650" containerID="a6045da0850d754dad86ca64641902f44db171a8cb4287c2892e11cc5b0b0394" exitCode=0 Nov 25 14:03:17 crc kubenswrapper[4702]: I1125 14:03:17.009016 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9" event={"ID":"401b6536-9504-448e-8acb-cb09c0cb0650","Type":"ContainerDied","Data":"a6045da0850d754dad86ca64641902f44db171a8cb4287c2892e11cc5b0b0394"} Nov 25 14:03:17 crc kubenswrapper[4702]: I1125 14:03:17.015739 4702 generic.go:334] "Generic (PLEG): container finished" podID="015eb990-26fa-467f-b645-fd3290426356" containerID="e9277dbdaed21c311649c2b542e9d9ae4cc59d3fc18c4b1114a4f74f3f196534" exitCode=0 Nov 25 14:03:17 crc kubenswrapper[4702]: I1125 14:03:17.015817 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp" event={"ID":"015eb990-26fa-467f-b645-fd3290426356","Type":"ContainerDied","Data":"e9277dbdaed21c311649c2b542e9d9ae4cc59d3fc18c4b1114a4f74f3f196534"} Nov 25 14:03:18 crc kubenswrapper[4702]: I1125 14:03:18.026744 4702 generic.go:334] "Generic (PLEG): container finished" podID="401b6536-9504-448e-8acb-cb09c0cb0650" containerID="397e4de61b13fbf293fd86e8f9b70bc76a25224919288f50650f2629b342bdb6" exitCode=0 Nov 25 14:03:18 crc kubenswrapper[4702]: I1125 14:03:18.026816 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9" event={"ID":"401b6536-9504-448e-8acb-cb09c0cb0650","Type":"ContainerDied","Data":"397e4de61b13fbf293fd86e8f9b70bc76a25224919288f50650f2629b342bdb6"} Nov 25 14:03:18 crc kubenswrapper[4702]: I1125 14:03:18.291507 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp" Nov 25 14:03:18 crc kubenswrapper[4702]: I1125 14:03:18.392671 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngxqg\" (UniqueName: \"kubernetes.io/projected/015eb990-26fa-467f-b645-fd3290426356-kube-api-access-ngxqg\") pod \"015eb990-26fa-467f-b645-fd3290426356\" (UID: \"015eb990-26fa-467f-b645-fd3290426356\") " Nov 25 14:03:18 crc kubenswrapper[4702]: I1125 14:03:18.392794 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/015eb990-26fa-467f-b645-fd3290426356-util\") pod \"015eb990-26fa-467f-b645-fd3290426356\" (UID: \"015eb990-26fa-467f-b645-fd3290426356\") " Nov 25 14:03:18 crc kubenswrapper[4702]: I1125 14:03:18.392880 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/015eb990-26fa-467f-b645-fd3290426356-bundle\") pod \"015eb990-26fa-467f-b645-fd3290426356\" (UID: \"015eb990-26fa-467f-b645-fd3290426356\") " Nov 25 14:03:18 crc kubenswrapper[4702]: I1125 14:03:18.393951 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/015eb990-26fa-467f-b645-fd3290426356-bundle" (OuterVolumeSpecName: "bundle") pod "015eb990-26fa-467f-b645-fd3290426356" (UID: "015eb990-26fa-467f-b645-fd3290426356"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:03:18 crc kubenswrapper[4702]: I1125 14:03:18.401419 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/015eb990-26fa-467f-b645-fd3290426356-kube-api-access-ngxqg" (OuterVolumeSpecName: "kube-api-access-ngxqg") pod "015eb990-26fa-467f-b645-fd3290426356" (UID: "015eb990-26fa-467f-b645-fd3290426356"). InnerVolumeSpecName "kube-api-access-ngxqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:03:18 crc kubenswrapper[4702]: I1125 14:03:18.404346 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/015eb990-26fa-467f-b645-fd3290426356-util" (OuterVolumeSpecName: "util") pod "015eb990-26fa-467f-b645-fd3290426356" (UID: "015eb990-26fa-467f-b645-fd3290426356"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:03:18 crc kubenswrapper[4702]: I1125 14:03:18.495041 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngxqg\" (UniqueName: \"kubernetes.io/projected/015eb990-26fa-467f-b645-fd3290426356-kube-api-access-ngxqg\") on node \"crc\" DevicePath \"\"" Nov 25 14:03:18 crc kubenswrapper[4702]: I1125 14:03:18.495089 4702 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/015eb990-26fa-467f-b645-fd3290426356-util\") on node \"crc\" DevicePath \"\"" Nov 25 14:03:18 crc kubenswrapper[4702]: I1125 14:03:18.495100 4702 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/015eb990-26fa-467f-b645-fd3290426356-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 14:03:19 crc kubenswrapper[4702]: I1125 14:03:19.038138 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp" event={"ID":"015eb990-26fa-467f-b645-fd3290426356","Type":"ContainerDied","Data":"8ff7a2bb3071b6b5ccaa589e5a3b17edf35c464c377d77d06ed0ec03d631be07"} Nov 25 14:03:19 crc kubenswrapper[4702]: I1125 14:03:19.038926 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ff7a2bb3071b6b5ccaa589e5a3b17edf35c464c377d77d06ed0ec03d631be07" Nov 25 14:03:19 crc kubenswrapper[4702]: I1125 14:03:19.038179 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp" Nov 25 14:03:19 crc kubenswrapper[4702]: I1125 14:03:19.327825 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9" Nov 25 14:03:19 crc kubenswrapper[4702]: I1125 14:03:19.409663 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/401b6536-9504-448e-8acb-cb09c0cb0650-bundle\") pod \"401b6536-9504-448e-8acb-cb09c0cb0650\" (UID: \"401b6536-9504-448e-8acb-cb09c0cb0650\") " Nov 25 14:03:19 crc kubenswrapper[4702]: I1125 14:03:19.409858 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/401b6536-9504-448e-8acb-cb09c0cb0650-util\") pod \"401b6536-9504-448e-8acb-cb09c0cb0650\" (UID: \"401b6536-9504-448e-8acb-cb09c0cb0650\") " Nov 25 14:03:19 crc kubenswrapper[4702]: I1125 14:03:19.409904 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbd68\" (UniqueName: \"kubernetes.io/projected/401b6536-9504-448e-8acb-cb09c0cb0650-kube-api-access-bbd68\") pod \"401b6536-9504-448e-8acb-cb09c0cb0650\" (UID: \"401b6536-9504-448e-8acb-cb09c0cb0650\") " Nov 25 14:03:19 crc kubenswrapper[4702]: I1125 14:03:19.411212 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/401b6536-9504-448e-8acb-cb09c0cb0650-bundle" (OuterVolumeSpecName: "bundle") pod "401b6536-9504-448e-8acb-cb09c0cb0650" (UID: "401b6536-9504-448e-8acb-cb09c0cb0650"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:03:19 crc kubenswrapper[4702]: I1125 14:03:19.414064 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/401b6536-9504-448e-8acb-cb09c0cb0650-kube-api-access-bbd68" (OuterVolumeSpecName: "kube-api-access-bbd68") pod "401b6536-9504-448e-8acb-cb09c0cb0650" (UID: "401b6536-9504-448e-8acb-cb09c0cb0650"). InnerVolumeSpecName "kube-api-access-bbd68". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:03:19 crc kubenswrapper[4702]: I1125 14:03:19.427627 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/401b6536-9504-448e-8acb-cb09c0cb0650-util" (OuterVolumeSpecName: "util") pod "401b6536-9504-448e-8acb-cb09c0cb0650" (UID: "401b6536-9504-448e-8acb-cb09c0cb0650"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:03:19 crc kubenswrapper[4702]: I1125 14:03:19.511884 4702 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/401b6536-9504-448e-8acb-cb09c0cb0650-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 14:03:19 crc kubenswrapper[4702]: I1125 14:03:19.511935 4702 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/401b6536-9504-448e-8acb-cb09c0cb0650-util\") on node \"crc\" DevicePath \"\"" Nov 25 14:03:19 crc kubenswrapper[4702]: I1125 14:03:19.511944 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbd68\" (UniqueName: \"kubernetes.io/projected/401b6536-9504-448e-8acb-cb09c0cb0650-kube-api-access-bbd68\") on node \"crc\" DevicePath \"\"" Nov 25 14:03:20 crc kubenswrapper[4702]: I1125 14:03:20.056479 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9" event={"ID":"401b6536-9504-448e-8acb-cb09c0cb0650","Type":"ContainerDied","Data":"fb151adb73a593fcee10fc70d7a09ca15ce479032ef169b502b18840e521cad0"} Nov 25 14:03:20 crc kubenswrapper[4702]: I1125 14:03:20.057033 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb151adb73a593fcee10fc70d7a09ca15ce479032ef169b502b18840e521cad0" Nov 25 14:03:20 crc kubenswrapper[4702]: I1125 14:03:20.056588 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9" Nov 25 14:03:24 crc kubenswrapper[4702]: I1125 14:03:24.086758 4702 generic.go:334] "Generic (PLEG): container finished" podID="3339fdec-c17b-4052-a472-7c6e19475fe4" containerID="ab70360378ad003c37db0e35718f2a0c9d8610dab59da2005e4185453011362e" exitCode=0 Nov 25 14:03:24 crc kubenswrapper[4702]: I1125 14:03:24.086897 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-5j9zq" event={"ID":"3339fdec-c17b-4052-a472-7c6e19475fe4","Type":"ContainerDied","Data":"ab70360378ad003c37db0e35718f2a0c9d8610dab59da2005e4185453011362e"} Nov 25 14:03:25 crc kubenswrapper[4702]: I1125 14:03:25.362057 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-5j9zq" Nov 25 14:03:25 crc kubenswrapper[4702]: I1125 14:03:25.409580 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slxsb\" (UniqueName: \"kubernetes.io/projected/3339fdec-c17b-4052-a472-7c6e19475fe4-kube-api-access-slxsb\") pod \"3339fdec-c17b-4052-a472-7c6e19475fe4\" (UID: \"3339fdec-c17b-4052-a472-7c6e19475fe4\") " Nov 25 14:03:25 crc kubenswrapper[4702]: I1125 14:03:25.409710 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3339fdec-c17b-4052-a472-7c6e19475fe4-config-data\") pod \"3339fdec-c17b-4052-a472-7c6e19475fe4\" (UID: \"3339fdec-c17b-4052-a472-7c6e19475fe4\") " Nov 25 14:03:25 crc kubenswrapper[4702]: I1125 14:03:25.420355 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3339fdec-c17b-4052-a472-7c6e19475fe4-kube-api-access-slxsb" (OuterVolumeSpecName: "kube-api-access-slxsb") pod "3339fdec-c17b-4052-a472-7c6e19475fe4" (UID: "3339fdec-c17b-4052-a472-7c6e19475fe4"). InnerVolumeSpecName "kube-api-access-slxsb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:03:25 crc kubenswrapper[4702]: I1125 14:03:25.451838 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3339fdec-c17b-4052-a472-7c6e19475fe4-config-data" (OuterVolumeSpecName: "config-data") pod "3339fdec-c17b-4052-a472-7c6e19475fe4" (UID: "3339fdec-c17b-4052-a472-7c6e19475fe4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:03:25 crc kubenswrapper[4702]: I1125 14:03:25.512834 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slxsb\" (UniqueName: \"kubernetes.io/projected/3339fdec-c17b-4052-a472-7c6e19475fe4-kube-api-access-slxsb\") on node \"crc\" DevicePath \"\"" Nov 25 14:03:25 crc kubenswrapper[4702]: I1125 14:03:25.512870 4702 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3339fdec-c17b-4052-a472-7c6e19475fe4-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.102002 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-5j9zq" event={"ID":"3339fdec-c17b-4052-a472-7c6e19475fe4","Type":"ContainerDied","Data":"f579e6a25e1a47ea270d84b9cb1bdce50ec12b8bbd3991528f7657aa4d592f7e"} Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.102042 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f579e6a25e1a47ea270d84b9cb1bdce50ec12b8bbd3991528f7657aa4d592f7e" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.102055 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-5j9zq" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.323321 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-88sbx"] Nov 25 14:03:26 crc kubenswrapper[4702]: E1125 14:03:26.323939 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="401b6536-9504-448e-8acb-cb09c0cb0650" containerName="util" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.323965 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="401b6536-9504-448e-8acb-cb09c0cb0650" containerName="util" Nov 25 14:03:26 crc kubenswrapper[4702]: E1125 14:03:26.323978 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="401b6536-9504-448e-8acb-cb09c0cb0650" containerName="extract" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.323986 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="401b6536-9504-448e-8acb-cb09c0cb0650" containerName="extract" Nov 25 14:03:26 crc kubenswrapper[4702]: E1125 14:03:26.324009 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="015eb990-26fa-467f-b645-fd3290426356" containerName="extract" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.324017 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="015eb990-26fa-467f-b645-fd3290426356" containerName="extract" Nov 25 14:03:26 crc kubenswrapper[4702]: E1125 14:03:26.324026 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="015eb990-26fa-467f-b645-fd3290426356" containerName="pull" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.324032 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="015eb990-26fa-467f-b645-fd3290426356" containerName="pull" Nov 25 14:03:26 crc kubenswrapper[4702]: E1125 14:03:26.324051 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3339fdec-c17b-4052-a472-7c6e19475fe4" containerName="keystone-db-sync" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.324058 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="3339fdec-c17b-4052-a472-7c6e19475fe4" containerName="keystone-db-sync" Nov 25 14:03:26 crc kubenswrapper[4702]: E1125 14:03:26.324081 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="401b6536-9504-448e-8acb-cb09c0cb0650" containerName="pull" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.324088 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="401b6536-9504-448e-8acb-cb09c0cb0650" containerName="pull" Nov 25 14:03:26 crc kubenswrapper[4702]: E1125 14:03:26.324100 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="015eb990-26fa-467f-b645-fd3290426356" containerName="util" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.324106 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="015eb990-26fa-467f-b645-fd3290426356" containerName="util" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.324346 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="3339fdec-c17b-4052-a472-7c6e19475fe4" containerName="keystone-db-sync" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.324361 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="015eb990-26fa-467f-b645-fd3290426356" containerName="extract" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.324369 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="401b6536-9504-448e-8acb-cb09c0cb0650" containerName="extract" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.325877 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-88sbx" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.329330 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.330446 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.331202 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.331439 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"osp-secret" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.332023 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-42ddw" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.367385 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-88sbx"] Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.438465 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12d991ff-0e51-4ed3-8675-b3c1123bbda0-config-data\") pod \"keystone-bootstrap-88sbx\" (UID: \"12d991ff-0e51-4ed3-8675-b3c1123bbda0\") " pod="glance-kuttl-tests/keystone-bootstrap-88sbx" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.438515 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxs25\" (UniqueName: \"kubernetes.io/projected/12d991ff-0e51-4ed3-8675-b3c1123bbda0-kube-api-access-nxs25\") pod \"keystone-bootstrap-88sbx\" (UID: \"12d991ff-0e51-4ed3-8675-b3c1123bbda0\") " pod="glance-kuttl-tests/keystone-bootstrap-88sbx" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.438554 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/12d991ff-0e51-4ed3-8675-b3c1123bbda0-credential-keys\") pod \"keystone-bootstrap-88sbx\" (UID: \"12d991ff-0e51-4ed3-8675-b3c1123bbda0\") " pod="glance-kuttl-tests/keystone-bootstrap-88sbx" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.438584 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/12d991ff-0e51-4ed3-8675-b3c1123bbda0-fernet-keys\") pod \"keystone-bootstrap-88sbx\" (UID: \"12d991ff-0e51-4ed3-8675-b3c1123bbda0\") " pod="glance-kuttl-tests/keystone-bootstrap-88sbx" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.438609 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12d991ff-0e51-4ed3-8675-b3c1123bbda0-scripts\") pod \"keystone-bootstrap-88sbx\" (UID: \"12d991ff-0e51-4ed3-8675-b3c1123bbda0\") " pod="glance-kuttl-tests/keystone-bootstrap-88sbx" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.539900 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/12d991ff-0e51-4ed3-8675-b3c1123bbda0-credential-keys\") pod \"keystone-bootstrap-88sbx\" (UID: \"12d991ff-0e51-4ed3-8675-b3c1123bbda0\") " pod="glance-kuttl-tests/keystone-bootstrap-88sbx" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.539965 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/12d991ff-0e51-4ed3-8675-b3c1123bbda0-fernet-keys\") pod \"keystone-bootstrap-88sbx\" (UID: \"12d991ff-0e51-4ed3-8675-b3c1123bbda0\") " pod="glance-kuttl-tests/keystone-bootstrap-88sbx" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.539993 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12d991ff-0e51-4ed3-8675-b3c1123bbda0-scripts\") pod \"keystone-bootstrap-88sbx\" (UID: \"12d991ff-0e51-4ed3-8675-b3c1123bbda0\") " pod="glance-kuttl-tests/keystone-bootstrap-88sbx" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.540050 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12d991ff-0e51-4ed3-8675-b3c1123bbda0-config-data\") pod \"keystone-bootstrap-88sbx\" (UID: \"12d991ff-0e51-4ed3-8675-b3c1123bbda0\") " pod="glance-kuttl-tests/keystone-bootstrap-88sbx" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.540391 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxs25\" (UniqueName: \"kubernetes.io/projected/12d991ff-0e51-4ed3-8675-b3c1123bbda0-kube-api-access-nxs25\") pod \"keystone-bootstrap-88sbx\" (UID: \"12d991ff-0e51-4ed3-8675-b3c1123bbda0\") " pod="glance-kuttl-tests/keystone-bootstrap-88sbx" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.545411 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/12d991ff-0e51-4ed3-8675-b3c1123bbda0-fernet-keys\") pod \"keystone-bootstrap-88sbx\" (UID: \"12d991ff-0e51-4ed3-8675-b3c1123bbda0\") " pod="glance-kuttl-tests/keystone-bootstrap-88sbx" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.554945 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/12d991ff-0e51-4ed3-8675-b3c1123bbda0-credential-keys\") pod \"keystone-bootstrap-88sbx\" (UID: \"12d991ff-0e51-4ed3-8675-b3c1123bbda0\") " pod="glance-kuttl-tests/keystone-bootstrap-88sbx" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.555172 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12d991ff-0e51-4ed3-8675-b3c1123bbda0-config-data\") pod \"keystone-bootstrap-88sbx\" (UID: \"12d991ff-0e51-4ed3-8675-b3c1123bbda0\") " pod="glance-kuttl-tests/keystone-bootstrap-88sbx" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.556558 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12d991ff-0e51-4ed3-8675-b3c1123bbda0-scripts\") pod \"keystone-bootstrap-88sbx\" (UID: \"12d991ff-0e51-4ed3-8675-b3c1123bbda0\") " pod="glance-kuttl-tests/keystone-bootstrap-88sbx" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.561469 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxs25\" (UniqueName: \"kubernetes.io/projected/12d991ff-0e51-4ed3-8675-b3c1123bbda0-kube-api-access-nxs25\") pod \"keystone-bootstrap-88sbx\" (UID: \"12d991ff-0e51-4ed3-8675-b3c1123bbda0\") " pod="glance-kuttl-tests/keystone-bootstrap-88sbx" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.669099 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-88sbx" Nov 25 14:03:26 crc kubenswrapper[4702]: I1125 14:03:26.901595 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-88sbx"] Nov 25 14:03:27 crc kubenswrapper[4702]: I1125 14:03:27.174079 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-88sbx" event={"ID":"12d991ff-0e51-4ed3-8675-b3c1123bbda0","Type":"ContainerStarted","Data":"3112282e6ea4f2f10dd9003521c9711e8ee572d8eb0a9ae8a3e98e9aa34b5888"} Nov 25 14:03:27 crc kubenswrapper[4702]: I1125 14:03:27.174137 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-88sbx" event={"ID":"12d991ff-0e51-4ed3-8675-b3c1123bbda0","Type":"ContainerStarted","Data":"72d63a01496a934ace8392ba32cf44e334040f3ff692e1dd6e1232ea2d544b0d"} Nov 25 14:03:31 crc kubenswrapper[4702]: I1125 14:03:31.205410 4702 generic.go:334] "Generic (PLEG): container finished" podID="12d991ff-0e51-4ed3-8675-b3c1123bbda0" containerID="3112282e6ea4f2f10dd9003521c9711e8ee572d8eb0a9ae8a3e98e9aa34b5888" exitCode=0 Nov 25 14:03:31 crc kubenswrapper[4702]: I1125 14:03:31.205521 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-88sbx" event={"ID":"12d991ff-0e51-4ed3-8675-b3c1123bbda0","Type":"ContainerDied","Data":"3112282e6ea4f2f10dd9003521c9711e8ee572d8eb0a9ae8a3e98e9aa34b5888"} Nov 25 14:03:32 crc kubenswrapper[4702]: I1125 14:03:32.534223 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-88sbx" Nov 25 14:03:32 crc kubenswrapper[4702]: I1125 14:03:32.629644 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/12d991ff-0e51-4ed3-8675-b3c1123bbda0-fernet-keys\") pod \"12d991ff-0e51-4ed3-8675-b3c1123bbda0\" (UID: \"12d991ff-0e51-4ed3-8675-b3c1123bbda0\") " Nov 25 14:03:32 crc kubenswrapper[4702]: I1125 14:03:32.629765 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12d991ff-0e51-4ed3-8675-b3c1123bbda0-config-data\") pod \"12d991ff-0e51-4ed3-8675-b3c1123bbda0\" (UID: \"12d991ff-0e51-4ed3-8675-b3c1123bbda0\") " Nov 25 14:03:32 crc kubenswrapper[4702]: I1125 14:03:32.629816 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12d991ff-0e51-4ed3-8675-b3c1123bbda0-scripts\") pod \"12d991ff-0e51-4ed3-8675-b3c1123bbda0\" (UID: \"12d991ff-0e51-4ed3-8675-b3c1123bbda0\") " Nov 25 14:03:32 crc kubenswrapper[4702]: I1125 14:03:32.629858 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxs25\" (UniqueName: \"kubernetes.io/projected/12d991ff-0e51-4ed3-8675-b3c1123bbda0-kube-api-access-nxs25\") pod \"12d991ff-0e51-4ed3-8675-b3c1123bbda0\" (UID: \"12d991ff-0e51-4ed3-8675-b3c1123bbda0\") " Nov 25 14:03:32 crc kubenswrapper[4702]: I1125 14:03:32.629991 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/12d991ff-0e51-4ed3-8675-b3c1123bbda0-credential-keys\") pod \"12d991ff-0e51-4ed3-8675-b3c1123bbda0\" (UID: \"12d991ff-0e51-4ed3-8675-b3c1123bbda0\") " Nov 25 14:03:32 crc kubenswrapper[4702]: I1125 14:03:32.644569 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12d991ff-0e51-4ed3-8675-b3c1123bbda0-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "12d991ff-0e51-4ed3-8675-b3c1123bbda0" (UID: "12d991ff-0e51-4ed3-8675-b3c1123bbda0"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:03:32 crc kubenswrapper[4702]: I1125 14:03:32.652800 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12d991ff-0e51-4ed3-8675-b3c1123bbda0-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "12d991ff-0e51-4ed3-8675-b3c1123bbda0" (UID: "12d991ff-0e51-4ed3-8675-b3c1123bbda0"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:03:32 crc kubenswrapper[4702]: I1125 14:03:32.652931 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12d991ff-0e51-4ed3-8675-b3c1123bbda0-scripts" (OuterVolumeSpecName: "scripts") pod "12d991ff-0e51-4ed3-8675-b3c1123bbda0" (UID: "12d991ff-0e51-4ed3-8675-b3c1123bbda0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:03:32 crc kubenswrapper[4702]: I1125 14:03:32.652967 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12d991ff-0e51-4ed3-8675-b3c1123bbda0-kube-api-access-nxs25" (OuterVolumeSpecName: "kube-api-access-nxs25") pod "12d991ff-0e51-4ed3-8675-b3c1123bbda0" (UID: "12d991ff-0e51-4ed3-8675-b3c1123bbda0"). InnerVolumeSpecName "kube-api-access-nxs25". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:03:32 crc kubenswrapper[4702]: I1125 14:03:32.676226 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12d991ff-0e51-4ed3-8675-b3c1123bbda0-config-data" (OuterVolumeSpecName: "config-data") pod "12d991ff-0e51-4ed3-8675-b3c1123bbda0" (UID: "12d991ff-0e51-4ed3-8675-b3c1123bbda0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:03:32 crc kubenswrapper[4702]: I1125 14:03:32.731686 4702 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/12d991ff-0e51-4ed3-8675-b3c1123bbda0-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 25 14:03:32 crc kubenswrapper[4702]: I1125 14:03:32.731725 4702 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12d991ff-0e51-4ed3-8675-b3c1123bbda0-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 14:03:32 crc kubenswrapper[4702]: I1125 14:03:32.731745 4702 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12d991ff-0e51-4ed3-8675-b3c1123bbda0-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 14:03:32 crc kubenswrapper[4702]: I1125 14:03:32.731758 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxs25\" (UniqueName: \"kubernetes.io/projected/12d991ff-0e51-4ed3-8675-b3c1123bbda0-kube-api-access-nxs25\") on node \"crc\" DevicePath \"\"" Nov 25 14:03:32 crc kubenswrapper[4702]: I1125 14:03:32.731769 4702 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/12d991ff-0e51-4ed3-8675-b3c1123bbda0-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 25 14:03:33 crc kubenswrapper[4702]: I1125 14:03:33.222050 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-88sbx" event={"ID":"12d991ff-0e51-4ed3-8675-b3c1123bbda0","Type":"ContainerDied","Data":"72d63a01496a934ace8392ba32cf44e334040f3ff692e1dd6e1232ea2d544b0d"} Nov 25 14:03:33 crc kubenswrapper[4702]: I1125 14:03:33.222427 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72d63a01496a934ace8392ba32cf44e334040f3ff692e1dd6e1232ea2d544b0d" Nov 25 14:03:33 crc kubenswrapper[4702]: I1125 14:03:33.222103 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-88sbx" Nov 25 14:03:33 crc kubenswrapper[4702]: I1125 14:03:33.325989 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-8c94d4449-cbvs8"] Nov 25 14:03:33 crc kubenswrapper[4702]: E1125 14:03:33.326285 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12d991ff-0e51-4ed3-8675-b3c1123bbda0" containerName="keystone-bootstrap" Nov 25 14:03:33 crc kubenswrapper[4702]: I1125 14:03:33.326301 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="12d991ff-0e51-4ed3-8675-b3c1123bbda0" containerName="keystone-bootstrap" Nov 25 14:03:33 crc kubenswrapper[4702]: I1125 14:03:33.326423 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="12d991ff-0e51-4ed3-8675-b3c1123bbda0" containerName="keystone-bootstrap" Nov 25 14:03:33 crc kubenswrapper[4702]: I1125 14:03:33.326899 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-8c94d4449-cbvs8" Nov 25 14:03:33 crc kubenswrapper[4702]: I1125 14:03:33.329138 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 25 14:03:33 crc kubenswrapper[4702]: I1125 14:03:33.329138 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 25 14:03:33 crc kubenswrapper[4702]: I1125 14:03:33.329318 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-42ddw" Nov 25 14:03:33 crc kubenswrapper[4702]: I1125 14:03:33.331622 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 25 14:03:33 crc kubenswrapper[4702]: I1125 14:03:33.416196 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-8c94d4449-cbvs8"] Nov 25 14:03:33 crc kubenswrapper[4702]: I1125 14:03:33.441480 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/791fc834-956f-4506-abf6-5d3695e23108-fernet-keys\") pod \"keystone-8c94d4449-cbvs8\" (UID: \"791fc834-956f-4506-abf6-5d3695e23108\") " pod="glance-kuttl-tests/keystone-8c94d4449-cbvs8" Nov 25 14:03:33 crc kubenswrapper[4702]: I1125 14:03:33.441539 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dg6jc\" (UniqueName: \"kubernetes.io/projected/791fc834-956f-4506-abf6-5d3695e23108-kube-api-access-dg6jc\") pod \"keystone-8c94d4449-cbvs8\" (UID: \"791fc834-956f-4506-abf6-5d3695e23108\") " pod="glance-kuttl-tests/keystone-8c94d4449-cbvs8" Nov 25 14:03:33 crc kubenswrapper[4702]: I1125 14:03:33.441567 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/791fc834-956f-4506-abf6-5d3695e23108-credential-keys\") pod \"keystone-8c94d4449-cbvs8\" (UID: \"791fc834-956f-4506-abf6-5d3695e23108\") " pod="glance-kuttl-tests/keystone-8c94d4449-cbvs8" Nov 25 14:03:33 crc kubenswrapper[4702]: I1125 14:03:33.441669 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/791fc834-956f-4506-abf6-5d3695e23108-scripts\") pod \"keystone-8c94d4449-cbvs8\" (UID: \"791fc834-956f-4506-abf6-5d3695e23108\") " pod="glance-kuttl-tests/keystone-8c94d4449-cbvs8" Nov 25 14:03:33 crc kubenswrapper[4702]: I1125 14:03:33.441741 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/791fc834-956f-4506-abf6-5d3695e23108-config-data\") pod \"keystone-8c94d4449-cbvs8\" (UID: \"791fc834-956f-4506-abf6-5d3695e23108\") " pod="glance-kuttl-tests/keystone-8c94d4449-cbvs8" Nov 25 14:03:33 crc kubenswrapper[4702]: I1125 14:03:33.542956 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/791fc834-956f-4506-abf6-5d3695e23108-scripts\") pod \"keystone-8c94d4449-cbvs8\" (UID: \"791fc834-956f-4506-abf6-5d3695e23108\") " pod="glance-kuttl-tests/keystone-8c94d4449-cbvs8" Nov 25 14:03:33 crc kubenswrapper[4702]: I1125 14:03:33.543021 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/791fc834-956f-4506-abf6-5d3695e23108-config-data\") pod \"keystone-8c94d4449-cbvs8\" (UID: \"791fc834-956f-4506-abf6-5d3695e23108\") " pod="glance-kuttl-tests/keystone-8c94d4449-cbvs8" Nov 25 14:03:33 crc kubenswrapper[4702]: I1125 14:03:33.543081 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/791fc834-956f-4506-abf6-5d3695e23108-fernet-keys\") pod \"keystone-8c94d4449-cbvs8\" (UID: \"791fc834-956f-4506-abf6-5d3695e23108\") " pod="glance-kuttl-tests/keystone-8c94d4449-cbvs8" Nov 25 14:03:33 crc kubenswrapper[4702]: I1125 14:03:33.543104 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dg6jc\" (UniqueName: \"kubernetes.io/projected/791fc834-956f-4506-abf6-5d3695e23108-kube-api-access-dg6jc\") pod \"keystone-8c94d4449-cbvs8\" (UID: \"791fc834-956f-4506-abf6-5d3695e23108\") " pod="glance-kuttl-tests/keystone-8c94d4449-cbvs8" Nov 25 14:03:33 crc kubenswrapper[4702]: I1125 14:03:33.543124 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/791fc834-956f-4506-abf6-5d3695e23108-credential-keys\") pod \"keystone-8c94d4449-cbvs8\" (UID: \"791fc834-956f-4506-abf6-5d3695e23108\") " pod="glance-kuttl-tests/keystone-8c94d4449-cbvs8" Nov 25 14:03:33 crc kubenswrapper[4702]: I1125 14:03:33.552126 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/791fc834-956f-4506-abf6-5d3695e23108-scripts\") pod \"keystone-8c94d4449-cbvs8\" (UID: \"791fc834-956f-4506-abf6-5d3695e23108\") " pod="glance-kuttl-tests/keystone-8c94d4449-cbvs8" Nov 25 14:03:33 crc kubenswrapper[4702]: I1125 14:03:33.552867 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/791fc834-956f-4506-abf6-5d3695e23108-fernet-keys\") pod \"keystone-8c94d4449-cbvs8\" (UID: \"791fc834-956f-4506-abf6-5d3695e23108\") " pod="glance-kuttl-tests/keystone-8c94d4449-cbvs8" Nov 25 14:03:33 crc kubenswrapper[4702]: I1125 14:03:33.555935 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/791fc834-956f-4506-abf6-5d3695e23108-config-data\") pod \"keystone-8c94d4449-cbvs8\" (UID: \"791fc834-956f-4506-abf6-5d3695e23108\") " pod="glance-kuttl-tests/keystone-8c94d4449-cbvs8" Nov 25 14:03:33 crc kubenswrapper[4702]: I1125 14:03:33.559795 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/791fc834-956f-4506-abf6-5d3695e23108-credential-keys\") pod \"keystone-8c94d4449-cbvs8\" (UID: \"791fc834-956f-4506-abf6-5d3695e23108\") " pod="glance-kuttl-tests/keystone-8c94d4449-cbvs8" Nov 25 14:03:33 crc kubenswrapper[4702]: I1125 14:03:33.585358 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dg6jc\" (UniqueName: \"kubernetes.io/projected/791fc834-956f-4506-abf6-5d3695e23108-kube-api-access-dg6jc\") pod \"keystone-8c94d4449-cbvs8\" (UID: \"791fc834-956f-4506-abf6-5d3695e23108\") " pod="glance-kuttl-tests/keystone-8c94d4449-cbvs8" Nov 25 14:03:33 crc kubenswrapper[4702]: I1125 14:03:33.642571 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-8c94d4449-cbvs8" Nov 25 14:03:33 crc kubenswrapper[4702]: I1125 14:03:33.964675 4702 patch_prober.go:28] interesting pod/machine-config-daemon-qqr76 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 14:03:33 crc kubenswrapper[4702]: I1125 14:03:33.964766 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 14:03:34 crc kubenswrapper[4702]: I1125 14:03:34.129908 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-8c94d4449-cbvs8"] Nov 25 14:03:34 crc kubenswrapper[4702]: W1125 14:03:34.137613 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod791fc834_956f_4506_abf6_5d3695e23108.slice/crio-035cb203e0ce9ff5182a928bcf14912ad42c6177a5e0e7a934d0a656c23ac795 WatchSource:0}: Error finding container 035cb203e0ce9ff5182a928bcf14912ad42c6177a5e0e7a934d0a656c23ac795: Status 404 returned error can't find the container with id 035cb203e0ce9ff5182a928bcf14912ad42c6177a5e0e7a934d0a656c23ac795 Nov 25 14:03:34 crc kubenswrapper[4702]: I1125 14:03:34.235363 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-8c94d4449-cbvs8" event={"ID":"791fc834-956f-4506-abf6-5d3695e23108","Type":"ContainerStarted","Data":"035cb203e0ce9ff5182a928bcf14912ad42c6177a5e0e7a934d0a656c23ac795"} Nov 25 14:03:35 crc kubenswrapper[4702]: I1125 14:03:35.263641 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-8c94d4449-cbvs8" event={"ID":"791fc834-956f-4506-abf6-5d3695e23108","Type":"ContainerStarted","Data":"5b2a086d0df55eb362fd898250269a28ce1d5563853fbcd94471d28295223bf2"} Nov 25 14:03:35 crc kubenswrapper[4702]: I1125 14:03:35.264031 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/keystone-8c94d4449-cbvs8" Nov 25 14:03:35 crc kubenswrapper[4702]: I1125 14:03:35.296470 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-8c94d4449-cbvs8" podStartSLOduration=2.29644995 podStartE2EDuration="2.29644995s" podCreationTimestamp="2025-11-25 14:03:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 14:03:35.293809885 +0000 UTC m=+1024.456460335" watchObservedRunningTime="2025-11-25 14:03:35.29644995 +0000 UTC m=+1024.459100400" Nov 25 14:03:35 crc kubenswrapper[4702]: I1125 14:03:35.397402 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-7b9f5ffc8f-6zsw2"] Nov 25 14:03:35 crc kubenswrapper[4702]: I1125 14:03:35.411644 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-7b9f5ffc8f-6zsw2" Nov 25 14:03:35 crc kubenswrapper[4702]: I1125 14:03:35.416646 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-service-cert" Nov 25 14:03:35 crc kubenswrapper[4702]: I1125 14:03:35.416704 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-dwx79" Nov 25 14:03:35 crc kubenswrapper[4702]: I1125 14:03:35.435582 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-7b9f5ffc8f-6zsw2"] Nov 25 14:03:35 crc kubenswrapper[4702]: I1125 14:03:35.469660 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/597dd90d-2ea6-4ea4-ad61-df59d4ca8944-webhook-cert\") pod \"swift-operator-controller-manager-7b9f5ffc8f-6zsw2\" (UID: \"597dd90d-2ea6-4ea4-ad61-df59d4ca8944\") " pod="openstack-operators/swift-operator-controller-manager-7b9f5ffc8f-6zsw2" Nov 25 14:03:35 crc kubenswrapper[4702]: I1125 14:03:35.469718 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/597dd90d-2ea6-4ea4-ad61-df59d4ca8944-apiservice-cert\") pod \"swift-operator-controller-manager-7b9f5ffc8f-6zsw2\" (UID: \"597dd90d-2ea6-4ea4-ad61-df59d4ca8944\") " pod="openstack-operators/swift-operator-controller-manager-7b9f5ffc8f-6zsw2" Nov 25 14:03:35 crc kubenswrapper[4702]: I1125 14:03:35.469742 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nr6xg\" (UniqueName: \"kubernetes.io/projected/597dd90d-2ea6-4ea4-ad61-df59d4ca8944-kube-api-access-nr6xg\") pod \"swift-operator-controller-manager-7b9f5ffc8f-6zsw2\" (UID: \"597dd90d-2ea6-4ea4-ad61-df59d4ca8944\") " pod="openstack-operators/swift-operator-controller-manager-7b9f5ffc8f-6zsw2" Nov 25 14:03:35 crc kubenswrapper[4702]: I1125 14:03:35.571375 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/597dd90d-2ea6-4ea4-ad61-df59d4ca8944-webhook-cert\") pod \"swift-operator-controller-manager-7b9f5ffc8f-6zsw2\" (UID: \"597dd90d-2ea6-4ea4-ad61-df59d4ca8944\") " pod="openstack-operators/swift-operator-controller-manager-7b9f5ffc8f-6zsw2" Nov 25 14:03:35 crc kubenswrapper[4702]: I1125 14:03:35.571452 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/597dd90d-2ea6-4ea4-ad61-df59d4ca8944-apiservice-cert\") pod \"swift-operator-controller-manager-7b9f5ffc8f-6zsw2\" (UID: \"597dd90d-2ea6-4ea4-ad61-df59d4ca8944\") " pod="openstack-operators/swift-operator-controller-manager-7b9f5ffc8f-6zsw2" Nov 25 14:03:35 crc kubenswrapper[4702]: I1125 14:03:35.571477 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nr6xg\" (UniqueName: \"kubernetes.io/projected/597dd90d-2ea6-4ea4-ad61-df59d4ca8944-kube-api-access-nr6xg\") pod \"swift-operator-controller-manager-7b9f5ffc8f-6zsw2\" (UID: \"597dd90d-2ea6-4ea4-ad61-df59d4ca8944\") " pod="openstack-operators/swift-operator-controller-manager-7b9f5ffc8f-6zsw2" Nov 25 14:03:35 crc kubenswrapper[4702]: I1125 14:03:35.576928 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/597dd90d-2ea6-4ea4-ad61-df59d4ca8944-apiservice-cert\") pod \"swift-operator-controller-manager-7b9f5ffc8f-6zsw2\" (UID: \"597dd90d-2ea6-4ea4-ad61-df59d4ca8944\") " pod="openstack-operators/swift-operator-controller-manager-7b9f5ffc8f-6zsw2" Nov 25 14:03:35 crc kubenswrapper[4702]: I1125 14:03:35.580177 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/597dd90d-2ea6-4ea4-ad61-df59d4ca8944-webhook-cert\") pod \"swift-operator-controller-manager-7b9f5ffc8f-6zsw2\" (UID: \"597dd90d-2ea6-4ea4-ad61-df59d4ca8944\") " pod="openstack-operators/swift-operator-controller-manager-7b9f5ffc8f-6zsw2" Nov 25 14:03:35 crc kubenswrapper[4702]: I1125 14:03:35.598122 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nr6xg\" (UniqueName: \"kubernetes.io/projected/597dd90d-2ea6-4ea4-ad61-df59d4ca8944-kube-api-access-nr6xg\") pod \"swift-operator-controller-manager-7b9f5ffc8f-6zsw2\" (UID: \"597dd90d-2ea6-4ea4-ad61-df59d4ca8944\") " pod="openstack-operators/swift-operator-controller-manager-7b9f5ffc8f-6zsw2" Nov 25 14:03:35 crc kubenswrapper[4702]: I1125 14:03:35.737265 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-7b9f5ffc8f-6zsw2" Nov 25 14:03:36 crc kubenswrapper[4702]: I1125 14:03:36.141680 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-7b9f5ffc8f-6zsw2"] Nov 25 14:03:36 crc kubenswrapper[4702]: W1125 14:03:36.145044 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod597dd90d_2ea6_4ea4_ad61_df59d4ca8944.slice/crio-2ae0d699a8cebdf8b6c25e101d428a1718a4802de137c94c37468cb57c4eb11f WatchSource:0}: Error finding container 2ae0d699a8cebdf8b6c25e101d428a1718a4802de137c94c37468cb57c4eb11f: Status 404 returned error can't find the container with id 2ae0d699a8cebdf8b6c25e101d428a1718a4802de137c94c37468cb57c4eb11f Nov 25 14:03:36 crc kubenswrapper[4702]: I1125 14:03:36.271216 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-7b9f5ffc8f-6zsw2" event={"ID":"597dd90d-2ea6-4ea4-ad61-df59d4ca8944","Type":"ContainerStarted","Data":"2ae0d699a8cebdf8b6c25e101d428a1718a4802de137c94c37468cb57c4eb11f"} Nov 25 14:03:39 crc kubenswrapper[4702]: I1125 14:03:39.294305 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-7b9f5ffc8f-6zsw2" event={"ID":"597dd90d-2ea6-4ea4-ad61-df59d4ca8944","Type":"ContainerStarted","Data":"e3e0d9d1bd5f46aa0540e742200f3a63ee78cece3f2a74fae1021383d88c6e02"} Nov 25 14:03:39 crc kubenswrapper[4702]: I1125 14:03:39.295373 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-7b9f5ffc8f-6zsw2" Nov 25 14:03:39 crc kubenswrapper[4702]: I1125 14:03:39.317808 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-7b9f5ffc8f-6zsw2" podStartSLOduration=2.014545782 podStartE2EDuration="4.317785707s" podCreationTimestamp="2025-11-25 14:03:35 +0000 UTC" firstStartedPulling="2025-11-25 14:03:36.14731035 +0000 UTC m=+1025.309960800" lastFinishedPulling="2025-11-25 14:03:38.450550275 +0000 UTC m=+1027.613200725" observedRunningTime="2025-11-25 14:03:39.313264932 +0000 UTC m=+1028.475915402" watchObservedRunningTime="2025-11-25 14:03:39.317785707 +0000 UTC m=+1028.480436157" Nov 25 14:03:44 crc kubenswrapper[4702]: I1125 14:03:44.361026 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-664bc6f655-7r9vd"] Nov 25 14:03:44 crc kubenswrapper[4702]: I1125 14:03:44.362291 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-664bc6f655-7r9vd" Nov 25 14:03:44 crc kubenswrapper[4702]: I1125 14:03:44.364793 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-wjl8j" Nov 25 14:03:44 crc kubenswrapper[4702]: I1125 14:03:44.365526 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-service-cert" Nov 25 14:03:44 crc kubenswrapper[4702]: I1125 14:03:44.373538 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-664bc6f655-7r9vd"] Nov 25 14:03:44 crc kubenswrapper[4702]: I1125 14:03:44.423094 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8611ff67-05cc-4795-83a1-a2be5b64dcd7-webhook-cert\") pod \"horizon-operator-controller-manager-664bc6f655-7r9vd\" (UID: \"8611ff67-05cc-4795-83a1-a2be5b64dcd7\") " pod="openstack-operators/horizon-operator-controller-manager-664bc6f655-7r9vd" Nov 25 14:03:44 crc kubenswrapper[4702]: I1125 14:03:44.423162 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8611ff67-05cc-4795-83a1-a2be5b64dcd7-apiservice-cert\") pod \"horizon-operator-controller-manager-664bc6f655-7r9vd\" (UID: \"8611ff67-05cc-4795-83a1-a2be5b64dcd7\") " pod="openstack-operators/horizon-operator-controller-manager-664bc6f655-7r9vd" Nov 25 14:03:44 crc kubenswrapper[4702]: I1125 14:03:44.423201 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b252z\" (UniqueName: \"kubernetes.io/projected/8611ff67-05cc-4795-83a1-a2be5b64dcd7-kube-api-access-b252z\") pod \"horizon-operator-controller-manager-664bc6f655-7r9vd\" (UID: \"8611ff67-05cc-4795-83a1-a2be5b64dcd7\") " pod="openstack-operators/horizon-operator-controller-manager-664bc6f655-7r9vd" Nov 25 14:03:44 crc kubenswrapper[4702]: I1125 14:03:44.525019 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8611ff67-05cc-4795-83a1-a2be5b64dcd7-apiservice-cert\") pod \"horizon-operator-controller-manager-664bc6f655-7r9vd\" (UID: \"8611ff67-05cc-4795-83a1-a2be5b64dcd7\") " pod="openstack-operators/horizon-operator-controller-manager-664bc6f655-7r9vd" Nov 25 14:03:44 crc kubenswrapper[4702]: I1125 14:03:44.526831 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b252z\" (UniqueName: \"kubernetes.io/projected/8611ff67-05cc-4795-83a1-a2be5b64dcd7-kube-api-access-b252z\") pod \"horizon-operator-controller-manager-664bc6f655-7r9vd\" (UID: \"8611ff67-05cc-4795-83a1-a2be5b64dcd7\") " pod="openstack-operators/horizon-operator-controller-manager-664bc6f655-7r9vd" Nov 25 14:03:44 crc kubenswrapper[4702]: I1125 14:03:44.527058 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8611ff67-05cc-4795-83a1-a2be5b64dcd7-webhook-cert\") pod \"horizon-operator-controller-manager-664bc6f655-7r9vd\" (UID: \"8611ff67-05cc-4795-83a1-a2be5b64dcd7\") " pod="openstack-operators/horizon-operator-controller-manager-664bc6f655-7r9vd" Nov 25 14:03:44 crc kubenswrapper[4702]: I1125 14:03:44.534294 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8611ff67-05cc-4795-83a1-a2be5b64dcd7-apiservice-cert\") pod \"horizon-operator-controller-manager-664bc6f655-7r9vd\" (UID: \"8611ff67-05cc-4795-83a1-a2be5b64dcd7\") " pod="openstack-operators/horizon-operator-controller-manager-664bc6f655-7r9vd" Nov 25 14:03:44 crc kubenswrapper[4702]: I1125 14:03:44.534341 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8611ff67-05cc-4795-83a1-a2be5b64dcd7-webhook-cert\") pod \"horizon-operator-controller-manager-664bc6f655-7r9vd\" (UID: \"8611ff67-05cc-4795-83a1-a2be5b64dcd7\") " pod="openstack-operators/horizon-operator-controller-manager-664bc6f655-7r9vd" Nov 25 14:03:44 crc kubenswrapper[4702]: I1125 14:03:44.547338 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b252z\" (UniqueName: \"kubernetes.io/projected/8611ff67-05cc-4795-83a1-a2be5b64dcd7-kube-api-access-b252z\") pod \"horizon-operator-controller-manager-664bc6f655-7r9vd\" (UID: \"8611ff67-05cc-4795-83a1-a2be5b64dcd7\") " pod="openstack-operators/horizon-operator-controller-manager-664bc6f655-7r9vd" Nov 25 14:03:44 crc kubenswrapper[4702]: I1125 14:03:44.689620 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-664bc6f655-7r9vd" Nov 25 14:03:44 crc kubenswrapper[4702]: I1125 14:03:44.951307 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-664bc6f655-7r9vd"] Nov 25 14:03:45 crc kubenswrapper[4702]: I1125 14:03:45.357784 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-664bc6f655-7r9vd" event={"ID":"8611ff67-05cc-4795-83a1-a2be5b64dcd7","Type":"ContainerStarted","Data":"f4b0df28b8257e2fc92c631936ea00650965eb1671ce988014b16c0b2a01349f"} Nov 25 14:03:45 crc kubenswrapper[4702]: I1125 14:03:45.742209 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-7b9f5ffc8f-6zsw2" Nov 25 14:03:47 crc kubenswrapper[4702]: I1125 14:03:47.380710 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-664bc6f655-7r9vd" event={"ID":"8611ff67-05cc-4795-83a1-a2be5b64dcd7","Type":"ContainerStarted","Data":"850bf229c0f3e66cb7814c8266fca4034812f3e721e64c9d4e6d372e8cfb3454"} Nov 25 14:03:47 crc kubenswrapper[4702]: I1125 14:03:47.381619 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-664bc6f655-7r9vd" Nov 25 14:03:47 crc kubenswrapper[4702]: I1125 14:03:47.402626 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-664bc6f655-7r9vd" podStartSLOduration=1.364769171 podStartE2EDuration="3.402596325s" podCreationTimestamp="2025-11-25 14:03:44 +0000 UTC" firstStartedPulling="2025-11-25 14:03:44.960564868 +0000 UTC m=+1034.123215318" lastFinishedPulling="2025-11-25 14:03:46.998392022 +0000 UTC m=+1036.161042472" observedRunningTime="2025-11-25 14:03:47.397746243 +0000 UTC m=+1036.560396713" watchObservedRunningTime="2025-11-25 14:03:47.402596325 +0000 UTC m=+1036.565246775" Nov 25 14:03:51 crc kubenswrapper[4702]: I1125 14:03:51.340685 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 25 14:03:51 crc kubenswrapper[4702]: I1125 14:03:51.347346 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Nov 25 14:03:51 crc kubenswrapper[4702]: I1125 14:03:51.350075 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-storage-config-data" Nov 25 14:03:51 crc kubenswrapper[4702]: I1125 14:03:51.350154 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-swift-dockercfg-7pzhs" Nov 25 14:03:51 crc kubenswrapper[4702]: I1125 14:03:51.350248 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-files" Nov 25 14:03:51 crc kubenswrapper[4702]: I1125 14:03:51.350305 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-conf" Nov 25 14:03:51 crc kubenswrapper[4702]: I1125 14:03:51.359558 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 25 14:03:51 crc kubenswrapper[4702]: I1125 14:03:51.448373 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f332564e-cc0f-4d05-8816-139528cf0bb0-etc-swift\") pod \"swift-storage-0\" (UID: \"f332564e-cc0f-4d05-8816-139528cf0bb0\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 14:03:51 crc kubenswrapper[4702]: I1125 14:03:51.448787 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/f332564e-cc0f-4d05-8816-139528cf0bb0-cache\") pod \"swift-storage-0\" (UID: \"f332564e-cc0f-4d05-8816-139528cf0bb0\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 14:03:51 crc kubenswrapper[4702]: I1125 14:03:51.448824 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/f332564e-cc0f-4d05-8816-139528cf0bb0-lock\") pod \"swift-storage-0\" (UID: \"f332564e-cc0f-4d05-8816-139528cf0bb0\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 14:03:51 crc kubenswrapper[4702]: I1125 14:03:51.449057 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98g9m\" (UniqueName: \"kubernetes.io/projected/f332564e-cc0f-4d05-8816-139528cf0bb0-kube-api-access-98g9m\") pod \"swift-storage-0\" (UID: \"f332564e-cc0f-4d05-8816-139528cf0bb0\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 14:03:51 crc kubenswrapper[4702]: I1125 14:03:51.449145 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"f332564e-cc0f-4d05-8816-139528cf0bb0\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 14:03:51 crc kubenswrapper[4702]: I1125 14:03:51.550941 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98g9m\" (UniqueName: \"kubernetes.io/projected/f332564e-cc0f-4d05-8816-139528cf0bb0-kube-api-access-98g9m\") pod \"swift-storage-0\" (UID: \"f332564e-cc0f-4d05-8816-139528cf0bb0\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 14:03:51 crc kubenswrapper[4702]: I1125 14:03:51.550999 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"f332564e-cc0f-4d05-8816-139528cf0bb0\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 14:03:51 crc kubenswrapper[4702]: I1125 14:03:51.551140 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f332564e-cc0f-4d05-8816-139528cf0bb0-etc-swift\") pod \"swift-storage-0\" (UID: \"f332564e-cc0f-4d05-8816-139528cf0bb0\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 14:03:51 crc kubenswrapper[4702]: I1125 14:03:51.551165 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/f332564e-cc0f-4d05-8816-139528cf0bb0-cache\") pod \"swift-storage-0\" (UID: \"f332564e-cc0f-4d05-8816-139528cf0bb0\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 14:03:51 crc kubenswrapper[4702]: I1125 14:03:51.551192 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/f332564e-cc0f-4d05-8816-139528cf0bb0-lock\") pod \"swift-storage-0\" (UID: \"f332564e-cc0f-4d05-8816-139528cf0bb0\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 14:03:51 crc kubenswrapper[4702]: E1125 14:03:51.551486 4702 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 14:03:51 crc kubenswrapper[4702]: E1125 14:03:51.551506 4702 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 25 14:03:51 crc kubenswrapper[4702]: E1125 14:03:51.551562 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f332564e-cc0f-4d05-8816-139528cf0bb0-etc-swift podName:f332564e-cc0f-4d05-8816-139528cf0bb0 nodeName:}" failed. No retries permitted until 2025-11-25 14:03:52.051544793 +0000 UTC m=+1041.214195243 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f332564e-cc0f-4d05-8816-139528cf0bb0-etc-swift") pod "swift-storage-0" (UID: "f332564e-cc0f-4d05-8816-139528cf0bb0") : configmap "swift-ring-files" not found Nov 25 14:03:51 crc kubenswrapper[4702]: I1125 14:03:51.551831 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/f332564e-cc0f-4d05-8816-139528cf0bb0-lock\") pod \"swift-storage-0\" (UID: \"f332564e-cc0f-4d05-8816-139528cf0bb0\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 14:03:51 crc kubenswrapper[4702]: I1125 14:03:51.551976 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"f332564e-cc0f-4d05-8816-139528cf0bb0\") device mount path \"/mnt/openstack/pv03\"" pod="glance-kuttl-tests/swift-storage-0" Nov 25 14:03:51 crc kubenswrapper[4702]: I1125 14:03:51.552125 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/f332564e-cc0f-4d05-8816-139528cf0bb0-cache\") pod \"swift-storage-0\" (UID: \"f332564e-cc0f-4d05-8816-139528cf0bb0\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 14:03:51 crc kubenswrapper[4702]: I1125 14:03:51.578708 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98g9m\" (UniqueName: \"kubernetes.io/projected/f332564e-cc0f-4d05-8816-139528cf0bb0-kube-api-access-98g9m\") pod \"swift-storage-0\" (UID: \"f332564e-cc0f-4d05-8816-139528cf0bb0\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 14:03:51 crc kubenswrapper[4702]: I1125 14:03:51.578964 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"f332564e-cc0f-4d05-8816-139528cf0bb0\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 14:03:51 crc kubenswrapper[4702]: I1125 14:03:51.966304 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-v2n5m"] Nov 25 14:03:51 crc kubenswrapper[4702]: I1125 14:03:51.967547 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-v2n5m" Nov 25 14:03:51 crc kubenswrapper[4702]: I1125 14:03:51.969455 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-config-data" Nov 25 14:03:51 crc kubenswrapper[4702]: I1125 14:03:51.969642 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-scripts" Nov 25 14:03:51 crc kubenswrapper[4702]: I1125 14:03:51.969453 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-proxy-config-data" Nov 25 14:03:51 crc kubenswrapper[4702]: I1125 14:03:51.981194 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-v2n5m"] Nov 25 14:03:51 crc kubenswrapper[4702]: I1125 14:03:51.997483 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-v2n5m"] Nov 25 14:03:51 crc kubenswrapper[4702]: E1125 14:03:51.998317 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[dispersionconf etc-swift kube-api-access-nq7qm ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[dispersionconf etc-swift kube-api-access-nq7qm ring-data-devices scripts swiftconf]: context canceled" pod="glance-kuttl-tests/swift-ring-rebalance-v2n5m" podUID="f4c05e81-ff2b-41e0-910d-480731d780d5" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.036951 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-k5rd6"] Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.038028 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-k5rd6" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.050181 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-k5rd6"] Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.061110 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f4c05e81-ff2b-41e0-910d-480731d780d5-etc-swift\") pod \"swift-ring-rebalance-v2n5m\" (UID: \"f4c05e81-ff2b-41e0-910d-480731d780d5\") " pod="glance-kuttl-tests/swift-ring-rebalance-v2n5m" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.061162 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f4c05e81-ff2b-41e0-910d-480731d780d5-scripts\") pod \"swift-ring-rebalance-v2n5m\" (UID: \"f4c05e81-ff2b-41e0-910d-480731d780d5\") " pod="glance-kuttl-tests/swift-ring-rebalance-v2n5m" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.061193 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nq7qm\" (UniqueName: \"kubernetes.io/projected/f4c05e81-ff2b-41e0-910d-480731d780d5-kube-api-access-nq7qm\") pod \"swift-ring-rebalance-v2n5m\" (UID: \"f4c05e81-ff2b-41e0-910d-480731d780d5\") " pod="glance-kuttl-tests/swift-ring-rebalance-v2n5m" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.061320 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f4c05e81-ff2b-41e0-910d-480731d780d5-ring-data-devices\") pod \"swift-ring-rebalance-v2n5m\" (UID: \"f4c05e81-ff2b-41e0-910d-480731d780d5\") " pod="glance-kuttl-tests/swift-ring-rebalance-v2n5m" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.061361 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f4c05e81-ff2b-41e0-910d-480731d780d5-dispersionconf\") pod \"swift-ring-rebalance-v2n5m\" (UID: \"f4c05e81-ff2b-41e0-910d-480731d780d5\") " pod="glance-kuttl-tests/swift-ring-rebalance-v2n5m" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.061399 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f4c05e81-ff2b-41e0-910d-480731d780d5-swiftconf\") pod \"swift-ring-rebalance-v2n5m\" (UID: \"f4c05e81-ff2b-41e0-910d-480731d780d5\") " pod="glance-kuttl-tests/swift-ring-rebalance-v2n5m" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.061439 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f332564e-cc0f-4d05-8816-139528cf0bb0-etc-swift\") pod \"swift-storage-0\" (UID: \"f332564e-cc0f-4d05-8816-139528cf0bb0\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 14:03:52 crc kubenswrapper[4702]: E1125 14:03:52.061606 4702 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 14:03:52 crc kubenswrapper[4702]: E1125 14:03:52.061622 4702 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 25 14:03:52 crc kubenswrapper[4702]: E1125 14:03:52.061670 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f332564e-cc0f-4d05-8816-139528cf0bb0-etc-swift podName:f332564e-cc0f-4d05-8816-139528cf0bb0 nodeName:}" failed. No retries permitted until 2025-11-25 14:03:53.061654192 +0000 UTC m=+1042.224304642 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f332564e-cc0f-4d05-8816-139528cf0bb0-etc-swift") pod "swift-storage-0" (UID: "f332564e-cc0f-4d05-8816-139528cf0bb0") : configmap "swift-ring-files" not found Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.162704 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0f80b097-f79c-47fb-832d-950dd6045b4e-ring-data-devices\") pod \"swift-ring-rebalance-k5rd6\" (UID: \"0f80b097-f79c-47fb-832d-950dd6045b4e\") " pod="glance-kuttl-tests/swift-ring-rebalance-k5rd6" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.162791 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0f80b097-f79c-47fb-832d-950dd6045b4e-etc-swift\") pod \"swift-ring-rebalance-k5rd6\" (UID: \"0f80b097-f79c-47fb-832d-950dd6045b4e\") " pod="glance-kuttl-tests/swift-ring-rebalance-k5rd6" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.162838 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f4c05e81-ff2b-41e0-910d-480731d780d5-etc-swift\") pod \"swift-ring-rebalance-v2n5m\" (UID: \"f4c05e81-ff2b-41e0-910d-480731d780d5\") " pod="glance-kuttl-tests/swift-ring-rebalance-v2n5m" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.162863 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f4c05e81-ff2b-41e0-910d-480731d780d5-scripts\") pod \"swift-ring-rebalance-v2n5m\" (UID: \"f4c05e81-ff2b-41e0-910d-480731d780d5\") " pod="glance-kuttl-tests/swift-ring-rebalance-v2n5m" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.162886 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0f80b097-f79c-47fb-832d-950dd6045b4e-swiftconf\") pod \"swift-ring-rebalance-k5rd6\" (UID: \"0f80b097-f79c-47fb-832d-950dd6045b4e\") " pod="glance-kuttl-tests/swift-ring-rebalance-k5rd6" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.162908 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nq7qm\" (UniqueName: \"kubernetes.io/projected/f4c05e81-ff2b-41e0-910d-480731d780d5-kube-api-access-nq7qm\") pod \"swift-ring-rebalance-v2n5m\" (UID: \"f4c05e81-ff2b-41e0-910d-480731d780d5\") " pod="glance-kuttl-tests/swift-ring-rebalance-v2n5m" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.162929 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f4c05e81-ff2b-41e0-910d-480731d780d5-ring-data-devices\") pod \"swift-ring-rebalance-v2n5m\" (UID: \"f4c05e81-ff2b-41e0-910d-480731d780d5\") " pod="glance-kuttl-tests/swift-ring-rebalance-v2n5m" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.162964 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f4c05e81-ff2b-41e0-910d-480731d780d5-dispersionconf\") pod \"swift-ring-rebalance-v2n5m\" (UID: \"f4c05e81-ff2b-41e0-910d-480731d780d5\") " pod="glance-kuttl-tests/swift-ring-rebalance-v2n5m" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.162998 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0f80b097-f79c-47fb-832d-950dd6045b4e-scripts\") pod \"swift-ring-rebalance-k5rd6\" (UID: \"0f80b097-f79c-47fb-832d-950dd6045b4e\") " pod="glance-kuttl-tests/swift-ring-rebalance-k5rd6" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.163022 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gd54k\" (UniqueName: \"kubernetes.io/projected/0f80b097-f79c-47fb-832d-950dd6045b4e-kube-api-access-gd54k\") pod \"swift-ring-rebalance-k5rd6\" (UID: \"0f80b097-f79c-47fb-832d-950dd6045b4e\") " pod="glance-kuttl-tests/swift-ring-rebalance-k5rd6" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.163055 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f4c05e81-ff2b-41e0-910d-480731d780d5-swiftconf\") pod \"swift-ring-rebalance-v2n5m\" (UID: \"f4c05e81-ff2b-41e0-910d-480731d780d5\") " pod="glance-kuttl-tests/swift-ring-rebalance-v2n5m" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.163118 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0f80b097-f79c-47fb-832d-950dd6045b4e-dispersionconf\") pod \"swift-ring-rebalance-k5rd6\" (UID: \"0f80b097-f79c-47fb-832d-950dd6045b4e\") " pod="glance-kuttl-tests/swift-ring-rebalance-k5rd6" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.163697 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f4c05e81-ff2b-41e0-910d-480731d780d5-etc-swift\") pod \"swift-ring-rebalance-v2n5m\" (UID: \"f4c05e81-ff2b-41e0-910d-480731d780d5\") " pod="glance-kuttl-tests/swift-ring-rebalance-v2n5m" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.164813 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f4c05e81-ff2b-41e0-910d-480731d780d5-scripts\") pod \"swift-ring-rebalance-v2n5m\" (UID: \"f4c05e81-ff2b-41e0-910d-480731d780d5\") " pod="glance-kuttl-tests/swift-ring-rebalance-v2n5m" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.164813 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f4c05e81-ff2b-41e0-910d-480731d780d5-ring-data-devices\") pod \"swift-ring-rebalance-v2n5m\" (UID: \"f4c05e81-ff2b-41e0-910d-480731d780d5\") " pod="glance-kuttl-tests/swift-ring-rebalance-v2n5m" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.178763 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f4c05e81-ff2b-41e0-910d-480731d780d5-dispersionconf\") pod \"swift-ring-rebalance-v2n5m\" (UID: \"f4c05e81-ff2b-41e0-910d-480731d780d5\") " pod="glance-kuttl-tests/swift-ring-rebalance-v2n5m" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.184738 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f4c05e81-ff2b-41e0-910d-480731d780d5-swiftconf\") pod \"swift-ring-rebalance-v2n5m\" (UID: \"f4c05e81-ff2b-41e0-910d-480731d780d5\") " pod="glance-kuttl-tests/swift-ring-rebalance-v2n5m" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.225527 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nq7qm\" (UniqueName: \"kubernetes.io/projected/f4c05e81-ff2b-41e0-910d-480731d780d5-kube-api-access-nq7qm\") pod \"swift-ring-rebalance-v2n5m\" (UID: \"f4c05e81-ff2b-41e0-910d-480731d780d5\") " pod="glance-kuttl-tests/swift-ring-rebalance-v2n5m" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.265357 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0f80b097-f79c-47fb-832d-950dd6045b4e-etc-swift\") pod \"swift-ring-rebalance-k5rd6\" (UID: \"0f80b097-f79c-47fb-832d-950dd6045b4e\") " pod="glance-kuttl-tests/swift-ring-rebalance-k5rd6" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.265763 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0f80b097-f79c-47fb-832d-950dd6045b4e-swiftconf\") pod \"swift-ring-rebalance-k5rd6\" (UID: \"0f80b097-f79c-47fb-832d-950dd6045b4e\") " pod="glance-kuttl-tests/swift-ring-rebalance-k5rd6" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.265968 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0f80b097-f79c-47fb-832d-950dd6045b4e-scripts\") pod \"swift-ring-rebalance-k5rd6\" (UID: \"0f80b097-f79c-47fb-832d-950dd6045b4e\") " pod="glance-kuttl-tests/swift-ring-rebalance-k5rd6" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.266103 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gd54k\" (UniqueName: \"kubernetes.io/projected/0f80b097-f79c-47fb-832d-950dd6045b4e-kube-api-access-gd54k\") pod \"swift-ring-rebalance-k5rd6\" (UID: \"0f80b097-f79c-47fb-832d-950dd6045b4e\") " pod="glance-kuttl-tests/swift-ring-rebalance-k5rd6" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.265954 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0f80b097-f79c-47fb-832d-950dd6045b4e-etc-swift\") pod \"swift-ring-rebalance-k5rd6\" (UID: \"0f80b097-f79c-47fb-832d-950dd6045b4e\") " pod="glance-kuttl-tests/swift-ring-rebalance-k5rd6" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.266450 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0f80b097-f79c-47fb-832d-950dd6045b4e-dispersionconf\") pod \"swift-ring-rebalance-k5rd6\" (UID: \"0f80b097-f79c-47fb-832d-950dd6045b4e\") " pod="glance-kuttl-tests/swift-ring-rebalance-k5rd6" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.266649 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0f80b097-f79c-47fb-832d-950dd6045b4e-ring-data-devices\") pod \"swift-ring-rebalance-k5rd6\" (UID: \"0f80b097-f79c-47fb-832d-950dd6045b4e\") " pod="glance-kuttl-tests/swift-ring-rebalance-k5rd6" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.266759 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0f80b097-f79c-47fb-832d-950dd6045b4e-scripts\") pod \"swift-ring-rebalance-k5rd6\" (UID: \"0f80b097-f79c-47fb-832d-950dd6045b4e\") " pod="glance-kuttl-tests/swift-ring-rebalance-k5rd6" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.267581 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0f80b097-f79c-47fb-832d-950dd6045b4e-ring-data-devices\") pod \"swift-ring-rebalance-k5rd6\" (UID: \"0f80b097-f79c-47fb-832d-950dd6045b4e\") " pod="glance-kuttl-tests/swift-ring-rebalance-k5rd6" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.273781 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0f80b097-f79c-47fb-832d-950dd6045b4e-swiftconf\") pod \"swift-ring-rebalance-k5rd6\" (UID: \"0f80b097-f79c-47fb-832d-950dd6045b4e\") " pod="glance-kuttl-tests/swift-ring-rebalance-k5rd6" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.273798 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0f80b097-f79c-47fb-832d-950dd6045b4e-dispersionconf\") pod \"swift-ring-rebalance-k5rd6\" (UID: \"0f80b097-f79c-47fb-832d-950dd6045b4e\") " pod="glance-kuttl-tests/swift-ring-rebalance-k5rd6" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.291689 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gd54k\" (UniqueName: \"kubernetes.io/projected/0f80b097-f79c-47fb-832d-950dd6045b4e-kube-api-access-gd54k\") pod \"swift-ring-rebalance-k5rd6\" (UID: \"0f80b097-f79c-47fb-832d-950dd6045b4e\") " pod="glance-kuttl-tests/swift-ring-rebalance-k5rd6" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.358698 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-k5rd6" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.417480 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-v2n5m" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.429294 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-v2n5m" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.470805 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f4c05e81-ff2b-41e0-910d-480731d780d5-swiftconf\") pod \"f4c05e81-ff2b-41e0-910d-480731d780d5\" (UID: \"f4c05e81-ff2b-41e0-910d-480731d780d5\") " Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.470847 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f4c05e81-ff2b-41e0-910d-480731d780d5-dispersionconf\") pod \"f4c05e81-ff2b-41e0-910d-480731d780d5\" (UID: \"f4c05e81-ff2b-41e0-910d-480731d780d5\") " Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.470896 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f4c05e81-ff2b-41e0-910d-480731d780d5-scripts\") pod \"f4c05e81-ff2b-41e0-910d-480731d780d5\" (UID: \"f4c05e81-ff2b-41e0-910d-480731d780d5\") " Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.470960 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nq7qm\" (UniqueName: \"kubernetes.io/projected/f4c05e81-ff2b-41e0-910d-480731d780d5-kube-api-access-nq7qm\") pod \"f4c05e81-ff2b-41e0-910d-480731d780d5\" (UID: \"f4c05e81-ff2b-41e0-910d-480731d780d5\") " Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.471001 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f4c05e81-ff2b-41e0-910d-480731d780d5-ring-data-devices\") pod \"f4c05e81-ff2b-41e0-910d-480731d780d5\" (UID: \"f4c05e81-ff2b-41e0-910d-480731d780d5\") " Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.471071 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f4c05e81-ff2b-41e0-910d-480731d780d5-etc-swift\") pod \"f4c05e81-ff2b-41e0-910d-480731d780d5\" (UID: \"f4c05e81-ff2b-41e0-910d-480731d780d5\") " Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.471620 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4c05e81-ff2b-41e0-910d-480731d780d5-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "f4c05e81-ff2b-41e0-910d-480731d780d5" (UID: "f4c05e81-ff2b-41e0-910d-480731d780d5"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.471861 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4c05e81-ff2b-41e0-910d-480731d780d5-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "f4c05e81-ff2b-41e0-910d-480731d780d5" (UID: "f4c05e81-ff2b-41e0-910d-480731d780d5"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.472405 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4c05e81-ff2b-41e0-910d-480731d780d5-scripts" (OuterVolumeSpecName: "scripts") pod "f4c05e81-ff2b-41e0-910d-480731d780d5" (UID: "f4c05e81-ff2b-41e0-910d-480731d780d5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.477026 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4c05e81-ff2b-41e0-910d-480731d780d5-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "f4c05e81-ff2b-41e0-910d-480731d780d5" (UID: "f4c05e81-ff2b-41e0-910d-480731d780d5"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.477043 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4c05e81-ff2b-41e0-910d-480731d780d5-kube-api-access-nq7qm" (OuterVolumeSpecName: "kube-api-access-nq7qm") pod "f4c05e81-ff2b-41e0-910d-480731d780d5" (UID: "f4c05e81-ff2b-41e0-910d-480731d780d5"). InnerVolumeSpecName "kube-api-access-nq7qm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.477418 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4c05e81-ff2b-41e0-910d-480731d780d5-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "f4c05e81-ff2b-41e0-910d-480731d780d5" (UID: "f4c05e81-ff2b-41e0-910d-480731d780d5"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.572391 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nq7qm\" (UniqueName: \"kubernetes.io/projected/f4c05e81-ff2b-41e0-910d-480731d780d5-kube-api-access-nq7qm\") on node \"crc\" DevicePath \"\"" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.572832 4702 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f4c05e81-ff2b-41e0-910d-480731d780d5-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.572847 4702 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f4c05e81-ff2b-41e0-910d-480731d780d5-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.572861 4702 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f4c05e81-ff2b-41e0-910d-480731d780d5-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.572873 4702 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f4c05e81-ff2b-41e0-910d-480731d780d5-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.572886 4702 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f4c05e81-ff2b-41e0-910d-480731d780d5-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.833474 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-k5rd6"] Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.991850 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-index-nxqqk"] Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.992763 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-nxqqk" Nov 25 14:03:52 crc kubenswrapper[4702]: I1125 14:03:52.994958 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-index-dockercfg-vv9wk" Nov 25 14:03:53 crc kubenswrapper[4702]: I1125 14:03:53.009038 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-nxqqk"] Nov 25 14:03:53 crc kubenswrapper[4702]: I1125 14:03:53.082387 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f332564e-cc0f-4d05-8816-139528cf0bb0-etc-swift\") pod \"swift-storage-0\" (UID: \"f332564e-cc0f-4d05-8816-139528cf0bb0\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 14:03:53 crc kubenswrapper[4702]: I1125 14:03:53.082489 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xvgk\" (UniqueName: \"kubernetes.io/projected/cc0a8640-1af2-47ac-a821-324511457b9d-kube-api-access-2xvgk\") pod \"glance-operator-index-nxqqk\" (UID: \"cc0a8640-1af2-47ac-a821-324511457b9d\") " pod="openstack-operators/glance-operator-index-nxqqk" Nov 25 14:03:53 crc kubenswrapper[4702]: E1125 14:03:53.082606 4702 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 14:03:53 crc kubenswrapper[4702]: E1125 14:03:53.082638 4702 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 25 14:03:53 crc kubenswrapper[4702]: E1125 14:03:53.082707 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f332564e-cc0f-4d05-8816-139528cf0bb0-etc-swift podName:f332564e-cc0f-4d05-8816-139528cf0bb0 nodeName:}" failed. No retries permitted until 2025-11-25 14:03:55.082685013 +0000 UTC m=+1044.245335463 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f332564e-cc0f-4d05-8816-139528cf0bb0-etc-swift") pod "swift-storage-0" (UID: "f332564e-cc0f-4d05-8816-139528cf0bb0") : configmap "swift-ring-files" not found Nov 25 14:03:53 crc kubenswrapper[4702]: I1125 14:03:53.184443 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xvgk\" (UniqueName: \"kubernetes.io/projected/cc0a8640-1af2-47ac-a821-324511457b9d-kube-api-access-2xvgk\") pod \"glance-operator-index-nxqqk\" (UID: \"cc0a8640-1af2-47ac-a821-324511457b9d\") " pod="openstack-operators/glance-operator-index-nxqqk" Nov 25 14:03:53 crc kubenswrapper[4702]: I1125 14:03:53.204554 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xvgk\" (UniqueName: \"kubernetes.io/projected/cc0a8640-1af2-47ac-a821-324511457b9d-kube-api-access-2xvgk\") pod \"glance-operator-index-nxqqk\" (UID: \"cc0a8640-1af2-47ac-a821-324511457b9d\") " pod="openstack-operators/glance-operator-index-nxqqk" Nov 25 14:03:53 crc kubenswrapper[4702]: I1125 14:03:53.317861 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-nxqqk" Nov 25 14:03:53 crc kubenswrapper[4702]: I1125 14:03:53.430352 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-v2n5m" Nov 25 14:03:53 crc kubenswrapper[4702]: I1125 14:03:53.430400 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-k5rd6" event={"ID":"0f80b097-f79c-47fb-832d-950dd6045b4e","Type":"ContainerStarted","Data":"0752d0ee5af55803ca2e4fc72346c9a853fb3aae8167e86014210384f670a6f1"} Nov 25 14:03:53 crc kubenswrapper[4702]: I1125 14:03:53.492404 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-v2n5m"] Nov 25 14:03:53 crc kubenswrapper[4702]: I1125 14:03:53.502058 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-v2n5m"] Nov 25 14:03:53 crc kubenswrapper[4702]: I1125 14:03:53.759411 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-nxqqk"] Nov 25 14:03:53 crc kubenswrapper[4702]: W1125 14:03:53.769962 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc0a8640_1af2_47ac_a821_324511457b9d.slice/crio-0fd82aa411d1ed9624c1a1e31be9e76afcd6ffeb442d71863987a9a0e6df2a69 WatchSource:0}: Error finding container 0fd82aa411d1ed9624c1a1e31be9e76afcd6ffeb442d71863987a9a0e6df2a69: Status 404 returned error can't find the container with id 0fd82aa411d1ed9624c1a1e31be9e76afcd6ffeb442d71863987a9a0e6df2a69 Nov 25 14:03:53 crc kubenswrapper[4702]: I1125 14:03:53.823112 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4c05e81-ff2b-41e0-910d-480731d780d5" path="/var/lib/kubelet/pods/f4c05e81-ff2b-41e0-910d-480731d780d5/volumes" Nov 25 14:03:54 crc kubenswrapper[4702]: I1125 14:03:54.440052 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-nxqqk" event={"ID":"cc0a8640-1af2-47ac-a821-324511457b9d","Type":"ContainerStarted","Data":"0fd82aa411d1ed9624c1a1e31be9e76afcd6ffeb442d71863987a9a0e6df2a69"} Nov 25 14:03:54 crc kubenswrapper[4702]: I1125 14:03:54.695613 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-664bc6f655-7r9vd" Nov 25 14:03:55 crc kubenswrapper[4702]: I1125 14:03:55.133902 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f332564e-cc0f-4d05-8816-139528cf0bb0-etc-swift\") pod \"swift-storage-0\" (UID: \"f332564e-cc0f-4d05-8816-139528cf0bb0\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 14:03:55 crc kubenswrapper[4702]: E1125 14:03:55.134171 4702 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 14:03:55 crc kubenswrapper[4702]: E1125 14:03:55.134194 4702 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 25 14:03:55 crc kubenswrapper[4702]: E1125 14:03:55.134277 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f332564e-cc0f-4d05-8816-139528cf0bb0-etc-swift podName:f332564e-cc0f-4d05-8816-139528cf0bb0 nodeName:}" failed. No retries permitted until 2025-11-25 14:03:59.134258493 +0000 UTC m=+1048.296909023 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f332564e-cc0f-4d05-8816-139528cf0bb0-etc-swift") pod "swift-storage-0" (UID: "f332564e-cc0f-4d05-8816-139528cf0bb0") : configmap "swift-ring-files" not found Nov 25 14:03:57 crc kubenswrapper[4702]: I1125 14:03:57.489593 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-k5rd6" event={"ID":"0f80b097-f79c-47fb-832d-950dd6045b4e","Type":"ContainerStarted","Data":"f86b9b2a30370cbe152d05fd411692c5b4c2ccf818aca6ad8c2c0996bbc4bdb7"} Nov 25 14:03:57 crc kubenswrapper[4702]: I1125 14:03:57.517007 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-ring-rebalance-k5rd6" podStartSLOduration=2.790882125 podStartE2EDuration="6.516976302s" podCreationTimestamp="2025-11-25 14:03:51 +0000 UTC" firstStartedPulling="2025-11-25 14:03:52.859508403 +0000 UTC m=+1042.022158853" lastFinishedPulling="2025-11-25 14:03:56.58560258 +0000 UTC m=+1045.748253030" observedRunningTime="2025-11-25 14:03:57.505172815 +0000 UTC m=+1046.667823265" watchObservedRunningTime="2025-11-25 14:03:57.516976302 +0000 UTC m=+1046.679626762" Nov 25 14:03:59 crc kubenswrapper[4702]: I1125 14:03:59.159613 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f332564e-cc0f-4d05-8816-139528cf0bb0-etc-swift\") pod \"swift-storage-0\" (UID: \"f332564e-cc0f-4d05-8816-139528cf0bb0\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 14:03:59 crc kubenswrapper[4702]: E1125 14:03:59.159845 4702 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 14:03:59 crc kubenswrapper[4702]: E1125 14:03:59.160149 4702 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 25 14:03:59 crc kubenswrapper[4702]: E1125 14:03:59.160218 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f332564e-cc0f-4d05-8816-139528cf0bb0-etc-swift podName:f332564e-cc0f-4d05-8816-139528cf0bb0 nodeName:}" failed. No retries permitted until 2025-11-25 14:04:07.160195644 +0000 UTC m=+1056.322846094 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f332564e-cc0f-4d05-8816-139528cf0bb0-etc-swift") pod "swift-storage-0" (UID: "f332564e-cc0f-4d05-8816-139528cf0bb0") : configmap "swift-ring-files" not found Nov 25 14:03:59 crc kubenswrapper[4702]: I1125 14:03:59.515839 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-nxqqk" event={"ID":"cc0a8640-1af2-47ac-a821-324511457b9d","Type":"ContainerStarted","Data":"ca3cec998bacfeb076d1f517dcad400dad0ec45487120ca4677bc4a4d780981a"} Nov 25 14:03:59 crc kubenswrapper[4702]: I1125 14:03:59.538349 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-index-nxqqk" podStartSLOduration=2.160685823 podStartE2EDuration="7.538213619s" podCreationTimestamp="2025-11-25 14:03:52 +0000 UTC" firstStartedPulling="2025-11-25 14:03:53.774507613 +0000 UTC m=+1042.937158063" lastFinishedPulling="2025-11-25 14:03:59.152035389 +0000 UTC m=+1048.314685859" observedRunningTime="2025-11-25 14:03:59.531415428 +0000 UTC m=+1048.694065888" watchObservedRunningTime="2025-11-25 14:03:59.538213619 +0000 UTC m=+1048.700864089" Nov 25 14:04:03 crc kubenswrapper[4702]: I1125 14:04:03.318726 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/glance-operator-index-nxqqk" Nov 25 14:04:03 crc kubenswrapper[4702]: I1125 14:04:03.319487 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-index-nxqqk" Nov 25 14:04:03 crc kubenswrapper[4702]: I1125 14:04:03.349231 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/glance-operator-index-nxqqk" Nov 25 14:04:03 crc kubenswrapper[4702]: I1125 14:04:03.963857 4702 patch_prober.go:28] interesting pod/machine-config-daemon-qqr76 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 14:04:03 crc kubenswrapper[4702]: I1125 14:04:03.963963 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 14:04:03 crc kubenswrapper[4702]: I1125 14:04:03.964044 4702 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" Nov 25 14:04:03 crc kubenswrapper[4702]: I1125 14:04:03.965105 4702 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"03504dcb763332c521878c39bb07b19a232e8fc711f4c606ea5726c5c5f4da1a"} pod="openshift-machine-config-operator/machine-config-daemon-qqr76" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 14:04:03 crc kubenswrapper[4702]: I1125 14:04:03.965197 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" containerID="cri-o://03504dcb763332c521878c39bb07b19a232e8fc711f4c606ea5726c5c5f4da1a" gracePeriod=600 Nov 25 14:04:04 crc kubenswrapper[4702]: I1125 14:04:04.554672 4702 generic.go:334] "Generic (PLEG): container finished" podID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerID="03504dcb763332c521878c39bb07b19a232e8fc711f4c606ea5726c5c5f4da1a" exitCode=0 Nov 25 14:04:04 crc kubenswrapper[4702]: I1125 14:04:04.554759 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" event={"ID":"3c491818-b0bb-4d82-a031-96e2dfac8c27","Type":"ContainerDied","Data":"03504dcb763332c521878c39bb07b19a232e8fc711f4c606ea5726c5c5f4da1a"} Nov 25 14:04:04 crc kubenswrapper[4702]: I1125 14:04:04.555297 4702 scope.go:117] "RemoveContainer" containerID="63ddfd96ea126ce49009797997159362938b731d35159663d233bd6f5c4e30d7" Nov 25 14:04:05 crc kubenswrapper[4702]: I1125 14:04:05.411568 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/keystone-8c94d4449-cbvs8" Nov 25 14:04:05 crc kubenswrapper[4702]: I1125 14:04:05.572799 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" event={"ID":"3c491818-b0bb-4d82-a031-96e2dfac8c27","Type":"ContainerStarted","Data":"de4c184eb16dc995f1ebd5c48e0561cca0a21907587a6c388b21643c8a15afa4"} Nov 25 14:04:05 crc kubenswrapper[4702]: I1125 14:04:05.575625 4702 generic.go:334] "Generic (PLEG): container finished" podID="0f80b097-f79c-47fb-832d-950dd6045b4e" containerID="f86b9b2a30370cbe152d05fd411692c5b4c2ccf818aca6ad8c2c0996bbc4bdb7" exitCode=0 Nov 25 14:04:05 crc kubenswrapper[4702]: I1125 14:04:05.575614 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-k5rd6" event={"ID":"0f80b097-f79c-47fb-832d-950dd6045b4e","Type":"ContainerDied","Data":"f86b9b2a30370cbe152d05fd411692c5b4c2ccf818aca6ad8c2c0996bbc4bdb7"} Nov 25 14:04:06 crc kubenswrapper[4702]: I1125 14:04:06.922197 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-k5rd6" Nov 25 14:04:07 crc kubenswrapper[4702]: I1125 14:04:07.122004 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0f80b097-f79c-47fb-832d-950dd6045b4e-etc-swift\") pod \"0f80b097-f79c-47fb-832d-950dd6045b4e\" (UID: \"0f80b097-f79c-47fb-832d-950dd6045b4e\") " Nov 25 14:04:07 crc kubenswrapper[4702]: I1125 14:04:07.122163 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0f80b097-f79c-47fb-832d-950dd6045b4e-scripts\") pod \"0f80b097-f79c-47fb-832d-950dd6045b4e\" (UID: \"0f80b097-f79c-47fb-832d-950dd6045b4e\") " Nov 25 14:04:07 crc kubenswrapper[4702]: I1125 14:04:07.122223 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0f80b097-f79c-47fb-832d-950dd6045b4e-swiftconf\") pod \"0f80b097-f79c-47fb-832d-950dd6045b4e\" (UID: \"0f80b097-f79c-47fb-832d-950dd6045b4e\") " Nov 25 14:04:07 crc kubenswrapper[4702]: I1125 14:04:07.122409 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gd54k\" (UniqueName: \"kubernetes.io/projected/0f80b097-f79c-47fb-832d-950dd6045b4e-kube-api-access-gd54k\") pod \"0f80b097-f79c-47fb-832d-950dd6045b4e\" (UID: \"0f80b097-f79c-47fb-832d-950dd6045b4e\") " Nov 25 14:04:07 crc kubenswrapper[4702]: I1125 14:04:07.122549 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0f80b097-f79c-47fb-832d-950dd6045b4e-ring-data-devices\") pod \"0f80b097-f79c-47fb-832d-950dd6045b4e\" (UID: \"0f80b097-f79c-47fb-832d-950dd6045b4e\") " Nov 25 14:04:07 crc kubenswrapper[4702]: I1125 14:04:07.122604 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0f80b097-f79c-47fb-832d-950dd6045b4e-dispersionconf\") pod \"0f80b097-f79c-47fb-832d-950dd6045b4e\" (UID: \"0f80b097-f79c-47fb-832d-950dd6045b4e\") " Nov 25 14:04:07 crc kubenswrapper[4702]: I1125 14:04:07.124373 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f80b097-f79c-47fb-832d-950dd6045b4e-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "0f80b097-f79c-47fb-832d-950dd6045b4e" (UID: "0f80b097-f79c-47fb-832d-950dd6045b4e"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 14:04:07 crc kubenswrapper[4702]: I1125 14:04:07.124429 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f80b097-f79c-47fb-832d-950dd6045b4e-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "0f80b097-f79c-47fb-832d-950dd6045b4e" (UID: "0f80b097-f79c-47fb-832d-950dd6045b4e"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:04:07 crc kubenswrapper[4702]: I1125 14:04:07.134325 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f80b097-f79c-47fb-832d-950dd6045b4e-kube-api-access-gd54k" (OuterVolumeSpecName: "kube-api-access-gd54k") pod "0f80b097-f79c-47fb-832d-950dd6045b4e" (UID: "0f80b097-f79c-47fb-832d-950dd6045b4e"). InnerVolumeSpecName "kube-api-access-gd54k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:04:07 crc kubenswrapper[4702]: I1125 14:04:07.135503 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f80b097-f79c-47fb-832d-950dd6045b4e-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "0f80b097-f79c-47fb-832d-950dd6045b4e" (UID: "0f80b097-f79c-47fb-832d-950dd6045b4e"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:04:07 crc kubenswrapper[4702]: I1125 14:04:07.145065 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f80b097-f79c-47fb-832d-950dd6045b4e-scripts" (OuterVolumeSpecName: "scripts") pod "0f80b097-f79c-47fb-832d-950dd6045b4e" (UID: "0f80b097-f79c-47fb-832d-950dd6045b4e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 14:04:07 crc kubenswrapper[4702]: I1125 14:04:07.156124 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f80b097-f79c-47fb-832d-950dd6045b4e-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "0f80b097-f79c-47fb-832d-950dd6045b4e" (UID: "0f80b097-f79c-47fb-832d-950dd6045b4e"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:04:07 crc kubenswrapper[4702]: I1125 14:04:07.225294 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f332564e-cc0f-4d05-8816-139528cf0bb0-etc-swift\") pod \"swift-storage-0\" (UID: \"f332564e-cc0f-4d05-8816-139528cf0bb0\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 14:04:07 crc kubenswrapper[4702]: I1125 14:04:07.225461 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gd54k\" (UniqueName: \"kubernetes.io/projected/0f80b097-f79c-47fb-832d-950dd6045b4e-kube-api-access-gd54k\") on node \"crc\" DevicePath \"\"" Nov 25 14:04:07 crc kubenswrapper[4702]: I1125 14:04:07.225477 4702 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0f80b097-f79c-47fb-832d-950dd6045b4e-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 25 14:04:07 crc kubenswrapper[4702]: I1125 14:04:07.225487 4702 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0f80b097-f79c-47fb-832d-950dd6045b4e-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 25 14:04:07 crc kubenswrapper[4702]: I1125 14:04:07.225497 4702 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0f80b097-f79c-47fb-832d-950dd6045b4e-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 25 14:04:07 crc kubenswrapper[4702]: I1125 14:04:07.225507 4702 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0f80b097-f79c-47fb-832d-950dd6045b4e-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 14:04:07 crc kubenswrapper[4702]: I1125 14:04:07.225516 4702 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0f80b097-f79c-47fb-832d-950dd6045b4e-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 25 14:04:07 crc kubenswrapper[4702]: I1125 14:04:07.231582 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f332564e-cc0f-4d05-8816-139528cf0bb0-etc-swift\") pod \"swift-storage-0\" (UID: \"f332564e-cc0f-4d05-8816-139528cf0bb0\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 14:04:07 crc kubenswrapper[4702]: I1125 14:04:07.269480 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Nov 25 14:04:07 crc kubenswrapper[4702]: I1125 14:04:07.594781 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-k5rd6" event={"ID":"0f80b097-f79c-47fb-832d-950dd6045b4e","Type":"ContainerDied","Data":"0752d0ee5af55803ca2e4fc72346c9a853fb3aae8167e86014210384f670a6f1"} Nov 25 14:04:07 crc kubenswrapper[4702]: I1125 14:04:07.594839 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0752d0ee5af55803ca2e4fc72346c9a853fb3aae8167e86014210384f670a6f1" Nov 25 14:04:07 crc kubenswrapper[4702]: I1125 14:04:07.594874 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-k5rd6" Nov 25 14:04:07 crc kubenswrapper[4702]: I1125 14:04:07.734061 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 25 14:04:07 crc kubenswrapper[4702]: W1125 14:04:07.734121 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf332564e_cc0f_4d05_8816_139528cf0bb0.slice/crio-71fb0da4512fe2497f99a05e3fc130c485e069b002fc0460c78edbdbc7047f15 WatchSource:0}: Error finding container 71fb0da4512fe2497f99a05e3fc130c485e069b002fc0460c78edbdbc7047f15: Status 404 returned error can't find the container with id 71fb0da4512fe2497f99a05e3fc130c485e069b002fc0460c78edbdbc7047f15 Nov 25 14:04:08 crc kubenswrapper[4702]: I1125 14:04:08.605603 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"f332564e-cc0f-4d05-8816-139528cf0bb0","Type":"ContainerStarted","Data":"71fb0da4512fe2497f99a05e3fc130c485e069b002fc0460c78edbdbc7047f15"} Nov 25 14:04:09 crc kubenswrapper[4702]: I1125 14:04:09.617367 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"f332564e-cc0f-4d05-8816-139528cf0bb0","Type":"ContainerStarted","Data":"050d9e36be2981a987f9d8d5cc9fb71e3da3ddcabaf1a7d42e7fc4e78a357548"} Nov 25 14:04:09 crc kubenswrapper[4702]: I1125 14:04:09.619781 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"f332564e-cc0f-4d05-8816-139528cf0bb0","Type":"ContainerStarted","Data":"586b83c5f7cfc2a89366739b1b5d8ac2b9267660df1d865955ad05c3bac270b7"} Nov 25 14:04:09 crc kubenswrapper[4702]: I1125 14:04:09.619891 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"f332564e-cc0f-4d05-8816-139528cf0bb0","Type":"ContainerStarted","Data":"6961d4c0c0bb922a306d2538ee58172d0b4c4ab4c7d6241187222296568c2e92"} Nov 25 14:04:09 crc kubenswrapper[4702]: I1125 14:04:09.619981 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"f332564e-cc0f-4d05-8816-139528cf0bb0","Type":"ContainerStarted","Data":"8a6eaab5dd2aa644816eeb3bd5bb35b516cc9fe358a3216316358fcd0c1b265d"} Nov 25 14:04:11 crc kubenswrapper[4702]: I1125 14:04:11.639191 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"f332564e-cc0f-4d05-8816-139528cf0bb0","Type":"ContainerStarted","Data":"a664449b6e926bd00d736b54248ae53b85d6badd3764bb86d9ceea557691c0c4"} Nov 25 14:04:11 crc kubenswrapper[4702]: I1125 14:04:11.640130 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"f332564e-cc0f-4d05-8816-139528cf0bb0","Type":"ContainerStarted","Data":"97046f4b6dd3838b30ea3b8570ac18c31602d339ea2068ae987934e4508d1b56"} Nov 25 14:04:11 crc kubenswrapper[4702]: I1125 14:04:11.640148 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"f332564e-cc0f-4d05-8816-139528cf0bb0","Type":"ContainerStarted","Data":"7ae3c982e67e56ed6b8d0a8a269959fca3a982730eda22a0104c22947a195701"} Nov 25 14:04:12 crc kubenswrapper[4702]: I1125 14:04:12.962459 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-k4rxf"] Nov 25 14:04:12 crc kubenswrapper[4702]: E1125 14:04:12.963463 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f80b097-f79c-47fb-832d-950dd6045b4e" containerName="swift-ring-rebalance" Nov 25 14:04:12 crc kubenswrapper[4702]: I1125 14:04:12.963487 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f80b097-f79c-47fb-832d-950dd6045b4e" containerName="swift-ring-rebalance" Nov 25 14:04:12 crc kubenswrapper[4702]: I1125 14:04:12.963669 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f80b097-f79c-47fb-832d-950dd6045b4e" containerName="swift-ring-rebalance" Nov 25 14:04:12 crc kubenswrapper[4702]: I1125 14:04:12.964681 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-k4rxf" Nov 25 14:04:12 crc kubenswrapper[4702]: I1125 14:04:12.966943 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-proxy-config-data" Nov 25 14:04:12 crc kubenswrapper[4702]: I1125 14:04:12.977084 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-k4rxf"] Nov 25 14:04:13 crc kubenswrapper[4702]: I1125 14:04:13.127384 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9nrl\" (UniqueName: \"kubernetes.io/projected/058da3b1-d848-4742-9454-f980618f76bd-kube-api-access-m9nrl\") pod \"swift-proxy-6bd58cfcf7-k4rxf\" (UID: \"058da3b1-d848-4742-9454-f980618f76bd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-k4rxf" Nov 25 14:04:13 crc kubenswrapper[4702]: I1125 14:04:13.127891 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/058da3b1-d848-4742-9454-f980618f76bd-log-httpd\") pod \"swift-proxy-6bd58cfcf7-k4rxf\" (UID: \"058da3b1-d848-4742-9454-f980618f76bd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-k4rxf" Nov 25 14:04:13 crc kubenswrapper[4702]: I1125 14:04:13.128012 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/058da3b1-d848-4742-9454-f980618f76bd-etc-swift\") pod \"swift-proxy-6bd58cfcf7-k4rxf\" (UID: \"058da3b1-d848-4742-9454-f980618f76bd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-k4rxf" Nov 25 14:04:13 crc kubenswrapper[4702]: I1125 14:04:13.128148 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/058da3b1-d848-4742-9454-f980618f76bd-run-httpd\") pod \"swift-proxy-6bd58cfcf7-k4rxf\" (UID: \"058da3b1-d848-4742-9454-f980618f76bd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-k4rxf" Nov 25 14:04:13 crc kubenswrapper[4702]: I1125 14:04:13.128294 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/058da3b1-d848-4742-9454-f980618f76bd-config-data\") pod \"swift-proxy-6bd58cfcf7-k4rxf\" (UID: \"058da3b1-d848-4742-9454-f980618f76bd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-k4rxf" Nov 25 14:04:13 crc kubenswrapper[4702]: I1125 14:04:13.230781 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9nrl\" (UniqueName: \"kubernetes.io/projected/058da3b1-d848-4742-9454-f980618f76bd-kube-api-access-m9nrl\") pod \"swift-proxy-6bd58cfcf7-k4rxf\" (UID: \"058da3b1-d848-4742-9454-f980618f76bd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-k4rxf" Nov 25 14:04:13 crc kubenswrapper[4702]: I1125 14:04:13.231321 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/058da3b1-d848-4742-9454-f980618f76bd-log-httpd\") pod \"swift-proxy-6bd58cfcf7-k4rxf\" (UID: \"058da3b1-d848-4742-9454-f980618f76bd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-k4rxf" Nov 25 14:04:13 crc kubenswrapper[4702]: I1125 14:04:13.231492 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/058da3b1-d848-4742-9454-f980618f76bd-etc-swift\") pod \"swift-proxy-6bd58cfcf7-k4rxf\" (UID: \"058da3b1-d848-4742-9454-f980618f76bd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-k4rxf" Nov 25 14:04:13 crc kubenswrapper[4702]: I1125 14:04:13.231661 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/058da3b1-d848-4742-9454-f980618f76bd-run-httpd\") pod \"swift-proxy-6bd58cfcf7-k4rxf\" (UID: \"058da3b1-d848-4742-9454-f980618f76bd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-k4rxf" Nov 25 14:04:13 crc kubenswrapper[4702]: I1125 14:04:13.231797 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/058da3b1-d848-4742-9454-f980618f76bd-config-data\") pod \"swift-proxy-6bd58cfcf7-k4rxf\" (UID: \"058da3b1-d848-4742-9454-f980618f76bd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-k4rxf" Nov 25 14:04:13 crc kubenswrapper[4702]: I1125 14:04:13.231855 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/058da3b1-d848-4742-9454-f980618f76bd-log-httpd\") pod \"swift-proxy-6bd58cfcf7-k4rxf\" (UID: \"058da3b1-d848-4742-9454-f980618f76bd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-k4rxf" Nov 25 14:04:13 crc kubenswrapper[4702]: I1125 14:04:13.232092 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/058da3b1-d848-4742-9454-f980618f76bd-run-httpd\") pod \"swift-proxy-6bd58cfcf7-k4rxf\" (UID: \"058da3b1-d848-4742-9454-f980618f76bd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-k4rxf" Nov 25 14:04:13 crc kubenswrapper[4702]: I1125 14:04:13.245771 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/058da3b1-d848-4742-9454-f980618f76bd-config-data\") pod \"swift-proxy-6bd58cfcf7-k4rxf\" (UID: \"058da3b1-d848-4742-9454-f980618f76bd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-k4rxf" Nov 25 14:04:13 crc kubenswrapper[4702]: I1125 14:04:13.246095 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/058da3b1-d848-4742-9454-f980618f76bd-etc-swift\") pod \"swift-proxy-6bd58cfcf7-k4rxf\" (UID: \"058da3b1-d848-4742-9454-f980618f76bd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-k4rxf" Nov 25 14:04:13 crc kubenswrapper[4702]: I1125 14:04:13.265138 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9nrl\" (UniqueName: \"kubernetes.io/projected/058da3b1-d848-4742-9454-f980618f76bd-kube-api-access-m9nrl\") pod \"swift-proxy-6bd58cfcf7-k4rxf\" (UID: \"058da3b1-d848-4742-9454-f980618f76bd\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-k4rxf" Nov 25 14:04:13 crc kubenswrapper[4702]: I1125 14:04:13.283039 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-k4rxf" Nov 25 14:04:13 crc kubenswrapper[4702]: I1125 14:04:13.352918 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-index-nxqqk" Nov 25 14:04:13 crc kubenswrapper[4702]: I1125 14:04:13.566517 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-k4rxf"] Nov 25 14:04:13 crc kubenswrapper[4702]: I1125 14:04:13.659340 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-k4rxf" event={"ID":"058da3b1-d848-4742-9454-f980618f76bd","Type":"ContainerStarted","Data":"fbb86d06bb93f722f4ad1625841fe77a536d10870a984a05c4f524f019508ca0"} Nov 25 14:04:14 crc kubenswrapper[4702]: I1125 14:04:14.668350 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-k4rxf" event={"ID":"058da3b1-d848-4742-9454-f980618f76bd","Type":"ContainerStarted","Data":"80c1ab3429979ad098efd2bada2d98a5189983d9fa0198c237dbd0d9c9faabc4"} Nov 25 14:04:14 crc kubenswrapper[4702]: I1125 14:04:14.668906 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-k4rxf" Nov 25 14:04:14 crc kubenswrapper[4702]: I1125 14:04:14.668919 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-k4rxf" event={"ID":"058da3b1-d848-4742-9454-f980618f76bd","Type":"ContainerStarted","Data":"41dc25905e9f6500e063636bc90482b815472a2d40a07938cb9101c5f7d8dd8d"} Nov 25 14:04:14 crc kubenswrapper[4702]: I1125 14:04:14.673941 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"f332564e-cc0f-4d05-8816-139528cf0bb0","Type":"ContainerStarted","Data":"83d7128b87e03b8088c036f551885f201c017f9ed344173688666341cbcede94"} Nov 25 14:04:15 crc kubenswrapper[4702]: I1125 14:04:15.256095 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-k4rxf" podStartSLOduration=3.256074153 podStartE2EDuration="3.256074153s" podCreationTimestamp="2025-11-25 14:04:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 14:04:14.69721815 +0000 UTC m=+1063.859868600" watchObservedRunningTime="2025-11-25 14:04:15.256074153 +0000 UTC m=+1064.418724603" Nov 25 14:04:15 crc kubenswrapper[4702]: I1125 14:04:15.260471 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p"] Nov 25 14:04:15 crc kubenswrapper[4702]: I1125 14:04:15.262295 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p" Nov 25 14:04:15 crc kubenswrapper[4702]: I1125 14:04:15.286026 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p"] Nov 25 14:04:15 crc kubenswrapper[4702]: I1125 14:04:15.340185 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-6r4jq" Nov 25 14:04:15 crc kubenswrapper[4702]: I1125 14:04:15.367164 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jcgw\" (UniqueName: \"kubernetes.io/projected/bb1af68d-a7e9-428b-adcd-2a5b38d18fc7-kube-api-access-8jcgw\") pod \"ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p\" (UID: \"bb1af68d-a7e9-428b-adcd-2a5b38d18fc7\") " pod="openstack-operators/ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p" Nov 25 14:04:15 crc kubenswrapper[4702]: I1125 14:04:15.367642 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb1af68d-a7e9-428b-adcd-2a5b38d18fc7-util\") pod \"ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p\" (UID: \"bb1af68d-a7e9-428b-adcd-2a5b38d18fc7\") " pod="openstack-operators/ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p" Nov 25 14:04:15 crc kubenswrapper[4702]: I1125 14:04:15.367809 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb1af68d-a7e9-428b-adcd-2a5b38d18fc7-bundle\") pod \"ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p\" (UID: \"bb1af68d-a7e9-428b-adcd-2a5b38d18fc7\") " pod="openstack-operators/ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p" Nov 25 14:04:15 crc kubenswrapper[4702]: I1125 14:04:15.469260 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb1af68d-a7e9-428b-adcd-2a5b38d18fc7-bundle\") pod \"ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p\" (UID: \"bb1af68d-a7e9-428b-adcd-2a5b38d18fc7\") " pod="openstack-operators/ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p" Nov 25 14:04:15 crc kubenswrapper[4702]: I1125 14:04:15.469608 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jcgw\" (UniqueName: \"kubernetes.io/projected/bb1af68d-a7e9-428b-adcd-2a5b38d18fc7-kube-api-access-8jcgw\") pod \"ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p\" (UID: \"bb1af68d-a7e9-428b-adcd-2a5b38d18fc7\") " pod="openstack-operators/ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p" Nov 25 14:04:15 crc kubenswrapper[4702]: I1125 14:04:15.469709 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb1af68d-a7e9-428b-adcd-2a5b38d18fc7-util\") pod \"ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p\" (UID: \"bb1af68d-a7e9-428b-adcd-2a5b38d18fc7\") " pod="openstack-operators/ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p" Nov 25 14:04:15 crc kubenswrapper[4702]: I1125 14:04:15.470335 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb1af68d-a7e9-428b-adcd-2a5b38d18fc7-util\") pod \"ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p\" (UID: \"bb1af68d-a7e9-428b-adcd-2a5b38d18fc7\") " pod="openstack-operators/ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p" Nov 25 14:04:15 crc kubenswrapper[4702]: I1125 14:04:15.470743 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb1af68d-a7e9-428b-adcd-2a5b38d18fc7-bundle\") pod \"ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p\" (UID: \"bb1af68d-a7e9-428b-adcd-2a5b38d18fc7\") " pod="openstack-operators/ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p" Nov 25 14:04:15 crc kubenswrapper[4702]: I1125 14:04:15.500500 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jcgw\" (UniqueName: \"kubernetes.io/projected/bb1af68d-a7e9-428b-adcd-2a5b38d18fc7-kube-api-access-8jcgw\") pod \"ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p\" (UID: \"bb1af68d-a7e9-428b-adcd-2a5b38d18fc7\") " pod="openstack-operators/ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p" Nov 25 14:04:15 crc kubenswrapper[4702]: I1125 14:04:15.650142 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p" Nov 25 14:04:15 crc kubenswrapper[4702]: I1125 14:04:15.681961 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-k4rxf" Nov 25 14:04:16 crc kubenswrapper[4702]: I1125 14:04:16.646347 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p"] Nov 25 14:04:16 crc kubenswrapper[4702]: W1125 14:04:16.650289 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb1af68d_a7e9_428b_adcd_2a5b38d18fc7.slice/crio-03ffdf36043f2b573efb4287a3b941565bf6a75d8037fee9a694904da4d2d8e4 WatchSource:0}: Error finding container 03ffdf36043f2b573efb4287a3b941565bf6a75d8037fee9a694904da4d2d8e4: Status 404 returned error can't find the container with id 03ffdf36043f2b573efb4287a3b941565bf6a75d8037fee9a694904da4d2d8e4 Nov 25 14:04:16 crc kubenswrapper[4702]: I1125 14:04:16.697575 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p" event={"ID":"bb1af68d-a7e9-428b-adcd-2a5b38d18fc7","Type":"ContainerStarted","Data":"03ffdf36043f2b573efb4287a3b941565bf6a75d8037fee9a694904da4d2d8e4"} Nov 25 14:04:16 crc kubenswrapper[4702]: I1125 14:04:16.704248 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"f332564e-cc0f-4d05-8816-139528cf0bb0","Type":"ContainerStarted","Data":"214f31d8099af45211fa69fe385dfe359d32e69017d355f53378102980022963"} Nov 25 14:04:16 crc kubenswrapper[4702]: I1125 14:04:16.704304 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"f332564e-cc0f-4d05-8816-139528cf0bb0","Type":"ContainerStarted","Data":"580d0ee4a7043555d46f287bb28ede89e9622b1c1428f6d3ae59d3b743f6dcc0"} Nov 25 14:04:17 crc kubenswrapper[4702]: I1125 14:04:17.713984 4702 generic.go:334] "Generic (PLEG): container finished" podID="bb1af68d-a7e9-428b-adcd-2a5b38d18fc7" containerID="79d9051c2f07388856b086dac639a58d9052713d2edbe853974307f8b9c7ec0b" exitCode=0 Nov 25 14:04:17 crc kubenswrapper[4702]: I1125 14:04:17.714083 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p" event={"ID":"bb1af68d-a7e9-428b-adcd-2a5b38d18fc7","Type":"ContainerDied","Data":"79d9051c2f07388856b086dac639a58d9052713d2edbe853974307f8b9c7ec0b"} Nov 25 14:04:17 crc kubenswrapper[4702]: I1125 14:04:17.724685 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"f332564e-cc0f-4d05-8816-139528cf0bb0","Type":"ContainerStarted","Data":"7528c3641f5821f164812590f28f71b8a1acad839ba41699ddf3a6ca3d071a76"} Nov 25 14:04:17 crc kubenswrapper[4702]: I1125 14:04:17.725022 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"f332564e-cc0f-4d05-8816-139528cf0bb0","Type":"ContainerStarted","Data":"da0e0554e44e7a4e00844567fb6e75f229f7c9281782d6ebedb259c3e1b232ad"} Nov 25 14:04:17 crc kubenswrapper[4702]: I1125 14:04:17.725048 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"f332564e-cc0f-4d05-8816-139528cf0bb0","Type":"ContainerStarted","Data":"c7cfa10b7eba710d37ca48e8f55cc0ea4964c1317f0e170d0aa7a2ae259df1cb"} Nov 25 14:04:17 crc kubenswrapper[4702]: I1125 14:04:17.725061 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"f332564e-cc0f-4d05-8816-139528cf0bb0","Type":"ContainerStarted","Data":"cddacf87c73def4e6f2ebd303d22e32b30a14689fe3fde8df6662e0a0db6c5c6"} Nov 25 14:04:17 crc kubenswrapper[4702]: I1125 14:04:17.725070 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"f332564e-cc0f-4d05-8816-139528cf0bb0","Type":"ContainerStarted","Data":"7306462061d60763ef8a31b533bd55edaab3f5bb96fbdb11e3560f7a8114a599"} Nov 25 14:04:17 crc kubenswrapper[4702]: I1125 14:04:17.790942 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-storage-0" podStartSLOduration=19.304658537 podStartE2EDuration="27.79092132s" podCreationTimestamp="2025-11-25 14:03:50 +0000 UTC" firstStartedPulling="2025-11-25 14:04:07.746014265 +0000 UTC m=+1056.908664715" lastFinishedPulling="2025-11-25 14:04:16.232277038 +0000 UTC m=+1065.394927498" observedRunningTime="2025-11-25 14:04:17.776808915 +0000 UTC m=+1066.939459355" watchObservedRunningTime="2025-11-25 14:04:17.79092132 +0000 UTC m=+1066.953571770" Nov 25 14:04:18 crc kubenswrapper[4702]: I1125 14:04:18.737556 4702 generic.go:334] "Generic (PLEG): container finished" podID="bb1af68d-a7e9-428b-adcd-2a5b38d18fc7" containerID="0be582e73eda5cfaa7d5ceecbe43c0262b8e3bbb7539ebdb3c36eddc118c53da" exitCode=0 Nov 25 14:04:18 crc kubenswrapper[4702]: I1125 14:04:18.737635 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p" event={"ID":"bb1af68d-a7e9-428b-adcd-2a5b38d18fc7","Type":"ContainerDied","Data":"0be582e73eda5cfaa7d5ceecbe43c0262b8e3bbb7539ebdb3c36eddc118c53da"} Nov 25 14:04:19 crc kubenswrapper[4702]: I1125 14:04:19.747982 4702 generic.go:334] "Generic (PLEG): container finished" podID="bb1af68d-a7e9-428b-adcd-2a5b38d18fc7" containerID="f002a45bb083908e6a6236df0412f9467a59c612bb3f01e1655a0d4fa82c45bd" exitCode=0 Nov 25 14:04:19 crc kubenswrapper[4702]: I1125 14:04:19.748069 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p" event={"ID":"bb1af68d-a7e9-428b-adcd-2a5b38d18fc7","Type":"ContainerDied","Data":"f002a45bb083908e6a6236df0412f9467a59c612bb3f01e1655a0d4fa82c45bd"} Nov 25 14:04:21 crc kubenswrapper[4702]: I1125 14:04:21.085590 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p" Nov 25 14:04:21 crc kubenswrapper[4702]: I1125 14:04:21.169342 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb1af68d-a7e9-428b-adcd-2a5b38d18fc7-util\") pod \"bb1af68d-a7e9-428b-adcd-2a5b38d18fc7\" (UID: \"bb1af68d-a7e9-428b-adcd-2a5b38d18fc7\") " Nov 25 14:04:21 crc kubenswrapper[4702]: I1125 14:04:21.169532 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb1af68d-a7e9-428b-adcd-2a5b38d18fc7-bundle\") pod \"bb1af68d-a7e9-428b-adcd-2a5b38d18fc7\" (UID: \"bb1af68d-a7e9-428b-adcd-2a5b38d18fc7\") " Nov 25 14:04:21 crc kubenswrapper[4702]: I1125 14:04:21.169648 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jcgw\" (UniqueName: \"kubernetes.io/projected/bb1af68d-a7e9-428b-adcd-2a5b38d18fc7-kube-api-access-8jcgw\") pod \"bb1af68d-a7e9-428b-adcd-2a5b38d18fc7\" (UID: \"bb1af68d-a7e9-428b-adcd-2a5b38d18fc7\") " Nov 25 14:04:21 crc kubenswrapper[4702]: I1125 14:04:21.172673 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb1af68d-a7e9-428b-adcd-2a5b38d18fc7-bundle" (OuterVolumeSpecName: "bundle") pod "bb1af68d-a7e9-428b-adcd-2a5b38d18fc7" (UID: "bb1af68d-a7e9-428b-adcd-2a5b38d18fc7"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:04:21 crc kubenswrapper[4702]: I1125 14:04:21.177437 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb1af68d-a7e9-428b-adcd-2a5b38d18fc7-kube-api-access-8jcgw" (OuterVolumeSpecName: "kube-api-access-8jcgw") pod "bb1af68d-a7e9-428b-adcd-2a5b38d18fc7" (UID: "bb1af68d-a7e9-428b-adcd-2a5b38d18fc7"). InnerVolumeSpecName "kube-api-access-8jcgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:04:21 crc kubenswrapper[4702]: I1125 14:04:21.185561 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb1af68d-a7e9-428b-adcd-2a5b38d18fc7-util" (OuterVolumeSpecName: "util") pod "bb1af68d-a7e9-428b-adcd-2a5b38d18fc7" (UID: "bb1af68d-a7e9-428b-adcd-2a5b38d18fc7"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:04:21 crc kubenswrapper[4702]: I1125 14:04:21.271040 4702 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb1af68d-a7e9-428b-adcd-2a5b38d18fc7-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 14:04:21 crc kubenswrapper[4702]: I1125 14:04:21.271095 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jcgw\" (UniqueName: \"kubernetes.io/projected/bb1af68d-a7e9-428b-adcd-2a5b38d18fc7-kube-api-access-8jcgw\") on node \"crc\" DevicePath \"\"" Nov 25 14:04:21 crc kubenswrapper[4702]: I1125 14:04:21.271109 4702 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb1af68d-a7e9-428b-adcd-2a5b38d18fc7-util\") on node \"crc\" DevicePath \"\"" Nov 25 14:04:21 crc kubenswrapper[4702]: I1125 14:04:21.768362 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p" event={"ID":"bb1af68d-a7e9-428b-adcd-2a5b38d18fc7","Type":"ContainerDied","Data":"03ffdf36043f2b573efb4287a3b941565bf6a75d8037fee9a694904da4d2d8e4"} Nov 25 14:04:21 crc kubenswrapper[4702]: I1125 14:04:21.768806 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03ffdf36043f2b573efb4287a3b941565bf6a75d8037fee9a694904da4d2d8e4" Nov 25 14:04:21 crc kubenswrapper[4702]: I1125 14:04:21.768437 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p" Nov 25 14:04:23 crc kubenswrapper[4702]: I1125 14:04:23.286056 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-k4rxf" Nov 25 14:04:23 crc kubenswrapper[4702]: I1125 14:04:23.287996 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-k4rxf" Nov 25 14:04:32 crc kubenswrapper[4702]: I1125 14:04:32.322065 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-7cdb5cbc57-h5pdv"] Nov 25 14:04:32 crc kubenswrapper[4702]: E1125 14:04:32.323515 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb1af68d-a7e9-428b-adcd-2a5b38d18fc7" containerName="pull" Nov 25 14:04:32 crc kubenswrapper[4702]: I1125 14:04:32.323535 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb1af68d-a7e9-428b-adcd-2a5b38d18fc7" containerName="pull" Nov 25 14:04:32 crc kubenswrapper[4702]: E1125 14:04:32.323556 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb1af68d-a7e9-428b-adcd-2a5b38d18fc7" containerName="extract" Nov 25 14:04:32 crc kubenswrapper[4702]: I1125 14:04:32.323564 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb1af68d-a7e9-428b-adcd-2a5b38d18fc7" containerName="extract" Nov 25 14:04:32 crc kubenswrapper[4702]: E1125 14:04:32.323577 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb1af68d-a7e9-428b-adcd-2a5b38d18fc7" containerName="util" Nov 25 14:04:32 crc kubenswrapper[4702]: I1125 14:04:32.323585 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb1af68d-a7e9-428b-adcd-2a5b38d18fc7" containerName="util" Nov 25 14:04:32 crc kubenswrapper[4702]: I1125 14:04:32.323791 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb1af68d-a7e9-428b-adcd-2a5b38d18fc7" containerName="extract" Nov 25 14:04:32 crc kubenswrapper[4702]: I1125 14:04:32.324431 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7cdb5cbc57-h5pdv" Nov 25 14:04:32 crc kubenswrapper[4702]: I1125 14:04:32.327963 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-zj624" Nov 25 14:04:32 crc kubenswrapper[4702]: I1125 14:04:32.328497 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-service-cert" Nov 25 14:04:32 crc kubenswrapper[4702]: I1125 14:04:32.363112 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7886ece7-80ed-4443-8e1e-522f0628421b-apiservice-cert\") pod \"glance-operator-controller-manager-7cdb5cbc57-h5pdv\" (UID: \"7886ece7-80ed-4443-8e1e-522f0628421b\") " pod="openstack-operators/glance-operator-controller-manager-7cdb5cbc57-h5pdv" Nov 25 14:04:32 crc kubenswrapper[4702]: I1125 14:04:32.364366 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7886ece7-80ed-4443-8e1e-522f0628421b-webhook-cert\") pod \"glance-operator-controller-manager-7cdb5cbc57-h5pdv\" (UID: \"7886ece7-80ed-4443-8e1e-522f0628421b\") " pod="openstack-operators/glance-operator-controller-manager-7cdb5cbc57-h5pdv" Nov 25 14:04:32 crc kubenswrapper[4702]: I1125 14:04:32.364440 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqdkt\" (UniqueName: \"kubernetes.io/projected/7886ece7-80ed-4443-8e1e-522f0628421b-kube-api-access-pqdkt\") pod \"glance-operator-controller-manager-7cdb5cbc57-h5pdv\" (UID: \"7886ece7-80ed-4443-8e1e-522f0628421b\") " pod="openstack-operators/glance-operator-controller-manager-7cdb5cbc57-h5pdv" Nov 25 14:04:32 crc kubenswrapper[4702]: I1125 14:04:32.383180 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7cdb5cbc57-h5pdv"] Nov 25 14:04:32 crc kubenswrapper[4702]: I1125 14:04:32.466586 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqdkt\" (UniqueName: \"kubernetes.io/projected/7886ece7-80ed-4443-8e1e-522f0628421b-kube-api-access-pqdkt\") pod \"glance-operator-controller-manager-7cdb5cbc57-h5pdv\" (UID: \"7886ece7-80ed-4443-8e1e-522f0628421b\") " pod="openstack-operators/glance-operator-controller-manager-7cdb5cbc57-h5pdv" Nov 25 14:04:32 crc kubenswrapper[4702]: I1125 14:04:32.466688 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7886ece7-80ed-4443-8e1e-522f0628421b-apiservice-cert\") pod \"glance-operator-controller-manager-7cdb5cbc57-h5pdv\" (UID: \"7886ece7-80ed-4443-8e1e-522f0628421b\") " pod="openstack-operators/glance-operator-controller-manager-7cdb5cbc57-h5pdv" Nov 25 14:04:32 crc kubenswrapper[4702]: I1125 14:04:32.466808 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7886ece7-80ed-4443-8e1e-522f0628421b-webhook-cert\") pod \"glance-operator-controller-manager-7cdb5cbc57-h5pdv\" (UID: \"7886ece7-80ed-4443-8e1e-522f0628421b\") " pod="openstack-operators/glance-operator-controller-manager-7cdb5cbc57-h5pdv" Nov 25 14:04:32 crc kubenswrapper[4702]: I1125 14:04:32.476185 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7886ece7-80ed-4443-8e1e-522f0628421b-webhook-cert\") pod \"glance-operator-controller-manager-7cdb5cbc57-h5pdv\" (UID: \"7886ece7-80ed-4443-8e1e-522f0628421b\") " pod="openstack-operators/glance-operator-controller-manager-7cdb5cbc57-h5pdv" Nov 25 14:04:32 crc kubenswrapper[4702]: I1125 14:04:32.482297 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7886ece7-80ed-4443-8e1e-522f0628421b-apiservice-cert\") pod \"glance-operator-controller-manager-7cdb5cbc57-h5pdv\" (UID: \"7886ece7-80ed-4443-8e1e-522f0628421b\") " pod="openstack-operators/glance-operator-controller-manager-7cdb5cbc57-h5pdv" Nov 25 14:04:32 crc kubenswrapper[4702]: I1125 14:04:32.487561 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqdkt\" (UniqueName: \"kubernetes.io/projected/7886ece7-80ed-4443-8e1e-522f0628421b-kube-api-access-pqdkt\") pod \"glance-operator-controller-manager-7cdb5cbc57-h5pdv\" (UID: \"7886ece7-80ed-4443-8e1e-522f0628421b\") " pod="openstack-operators/glance-operator-controller-manager-7cdb5cbc57-h5pdv" Nov 25 14:04:32 crc kubenswrapper[4702]: I1125 14:04:32.678510 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7cdb5cbc57-h5pdv" Nov 25 14:04:33 crc kubenswrapper[4702]: I1125 14:04:33.138318 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7cdb5cbc57-h5pdv"] Nov 25 14:04:33 crc kubenswrapper[4702]: W1125 14:04:33.143059 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7886ece7_80ed_4443_8e1e_522f0628421b.slice/crio-6c6b48e588fa7ea11694d69cd8cbef856ca930f9b8a64075f7a55a49d55f6bc9 WatchSource:0}: Error finding container 6c6b48e588fa7ea11694d69cd8cbef856ca930f9b8a64075f7a55a49d55f6bc9: Status 404 returned error can't find the container with id 6c6b48e588fa7ea11694d69cd8cbef856ca930f9b8a64075f7a55a49d55f6bc9 Nov 25 14:04:33 crc kubenswrapper[4702]: I1125 14:04:33.883370 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7cdb5cbc57-h5pdv" event={"ID":"7886ece7-80ed-4443-8e1e-522f0628421b","Type":"ContainerStarted","Data":"6c6b48e588fa7ea11694d69cd8cbef856ca930f9b8a64075f7a55a49d55f6bc9"} Nov 25 14:04:35 crc kubenswrapper[4702]: I1125 14:04:35.901888 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7cdb5cbc57-h5pdv" event={"ID":"7886ece7-80ed-4443-8e1e-522f0628421b","Type":"ContainerStarted","Data":"d2a540575597edbf5b4fb94c01a346567d4102741ef57264278a7d59f06461b5"} Nov 25 14:04:35 crc kubenswrapper[4702]: I1125 14:04:35.902525 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-7cdb5cbc57-h5pdv" Nov 25 14:04:35 crc kubenswrapper[4702]: I1125 14:04:35.934389 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-7cdb5cbc57-h5pdv" podStartSLOduration=1.743036501 podStartE2EDuration="3.934361157s" podCreationTimestamp="2025-11-25 14:04:32 +0000 UTC" firstStartedPulling="2025-11-25 14:04:33.145184999 +0000 UTC m=+1082.307835449" lastFinishedPulling="2025-11-25 14:04:35.336509645 +0000 UTC m=+1084.499160105" observedRunningTime="2025-11-25 14:04:35.925077373 +0000 UTC m=+1085.087727833" watchObservedRunningTime="2025-11-25 14:04:35.934361157 +0000 UTC m=+1085.097011627" Nov 25 14:04:42 crc kubenswrapper[4702]: I1125 14:04:42.683089 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-7cdb5cbc57-h5pdv" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.507927 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.509954 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.512500 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts-9db6gc427h" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.513520 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.513887 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"openstack-config-secret" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.514283 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"default-dockercfg-cc8rt" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.526669 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.557469 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-9t6bx"] Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.559360 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-9t6bx" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.568873 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-f364-account-create-update-xpwg6"] Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.570426 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-f364-account-create-update-xpwg6" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.576716 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-9t6bx"] Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.582732 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.586384 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-f364-account-create-update-xpwg6"] Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.607832 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ck7fh\" (UniqueName: \"kubernetes.io/projected/28c2c617-0d70-4223-a7b3-0043eb3b7d87-kube-api-access-ck7fh\") pod \"glance-f364-account-create-update-xpwg6\" (UID: \"28c2c617-0d70-4223-a7b3-0043eb3b7d87\") " pod="glance-kuttl-tests/glance-f364-account-create-update-xpwg6" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.607914 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a4aaf26d-004b-4fe6-9896-dd0530fdd040-openstack-config\") pod \"openstackclient\" (UID: \"a4aaf26d-004b-4fe6-9896-dd0530fdd040\") " pod="glance-kuttl-tests/openstackclient" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.607945 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5q9x\" (UniqueName: \"kubernetes.io/projected/30d08778-c048-4b5f-b0d5-d018e4af5b1f-kube-api-access-d5q9x\") pod \"glance-db-create-9t6bx\" (UID: \"30d08778-c048-4b5f-b0d5-d018e4af5b1f\") " pod="glance-kuttl-tests/glance-db-create-9t6bx" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.607972 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/28c2c617-0d70-4223-a7b3-0043eb3b7d87-operator-scripts\") pod \"glance-f364-account-create-update-xpwg6\" (UID: \"28c2c617-0d70-4223-a7b3-0043eb3b7d87\") " pod="glance-kuttl-tests/glance-f364-account-create-update-xpwg6" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.608000 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mz69s\" (UniqueName: \"kubernetes.io/projected/a4aaf26d-004b-4fe6-9896-dd0530fdd040-kube-api-access-mz69s\") pod \"openstackclient\" (UID: \"a4aaf26d-004b-4fe6-9896-dd0530fdd040\") " pod="glance-kuttl-tests/openstackclient" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.608019 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30d08778-c048-4b5f-b0d5-d018e4af5b1f-operator-scripts\") pod \"glance-db-create-9t6bx\" (UID: \"30d08778-c048-4b5f-b0d5-d018e4af5b1f\") " pod="glance-kuttl-tests/glance-db-create-9t6bx" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.608052 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/a4aaf26d-004b-4fe6-9896-dd0530fdd040-openstack-scripts\") pod \"openstackclient\" (UID: \"a4aaf26d-004b-4fe6-9896-dd0530fdd040\") " pod="glance-kuttl-tests/openstackclient" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.608079 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a4aaf26d-004b-4fe6-9896-dd0530fdd040-openstack-config-secret\") pod \"openstackclient\" (UID: \"a4aaf26d-004b-4fe6-9896-dd0530fdd040\") " pod="glance-kuttl-tests/openstackclient" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.710756 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ck7fh\" (UniqueName: \"kubernetes.io/projected/28c2c617-0d70-4223-a7b3-0043eb3b7d87-kube-api-access-ck7fh\") pod \"glance-f364-account-create-update-xpwg6\" (UID: \"28c2c617-0d70-4223-a7b3-0043eb3b7d87\") " pod="glance-kuttl-tests/glance-f364-account-create-update-xpwg6" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.710873 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a4aaf26d-004b-4fe6-9896-dd0530fdd040-openstack-config\") pod \"openstackclient\" (UID: \"a4aaf26d-004b-4fe6-9896-dd0530fdd040\") " pod="glance-kuttl-tests/openstackclient" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.710909 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5q9x\" (UniqueName: \"kubernetes.io/projected/30d08778-c048-4b5f-b0d5-d018e4af5b1f-kube-api-access-d5q9x\") pod \"glance-db-create-9t6bx\" (UID: \"30d08778-c048-4b5f-b0d5-d018e4af5b1f\") " pod="glance-kuttl-tests/glance-db-create-9t6bx" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.710955 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/28c2c617-0d70-4223-a7b3-0043eb3b7d87-operator-scripts\") pod \"glance-f364-account-create-update-xpwg6\" (UID: \"28c2c617-0d70-4223-a7b3-0043eb3b7d87\") " pod="glance-kuttl-tests/glance-f364-account-create-update-xpwg6" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.710987 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mz69s\" (UniqueName: \"kubernetes.io/projected/a4aaf26d-004b-4fe6-9896-dd0530fdd040-kube-api-access-mz69s\") pod \"openstackclient\" (UID: \"a4aaf26d-004b-4fe6-9896-dd0530fdd040\") " pod="glance-kuttl-tests/openstackclient" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.711054 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30d08778-c048-4b5f-b0d5-d018e4af5b1f-operator-scripts\") pod \"glance-db-create-9t6bx\" (UID: \"30d08778-c048-4b5f-b0d5-d018e4af5b1f\") " pod="glance-kuttl-tests/glance-db-create-9t6bx" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.711086 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/a4aaf26d-004b-4fe6-9896-dd0530fdd040-openstack-scripts\") pod \"openstackclient\" (UID: \"a4aaf26d-004b-4fe6-9896-dd0530fdd040\") " pod="glance-kuttl-tests/openstackclient" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.711103 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a4aaf26d-004b-4fe6-9896-dd0530fdd040-openstack-config-secret\") pod \"openstackclient\" (UID: \"a4aaf26d-004b-4fe6-9896-dd0530fdd040\") " pod="glance-kuttl-tests/openstackclient" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.712141 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a4aaf26d-004b-4fe6-9896-dd0530fdd040-openstack-config\") pod \"openstackclient\" (UID: \"a4aaf26d-004b-4fe6-9896-dd0530fdd040\") " pod="glance-kuttl-tests/openstackclient" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.712373 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/28c2c617-0d70-4223-a7b3-0043eb3b7d87-operator-scripts\") pod \"glance-f364-account-create-update-xpwg6\" (UID: \"28c2c617-0d70-4223-a7b3-0043eb3b7d87\") " pod="glance-kuttl-tests/glance-f364-account-create-update-xpwg6" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.712466 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30d08778-c048-4b5f-b0d5-d018e4af5b1f-operator-scripts\") pod \"glance-db-create-9t6bx\" (UID: \"30d08778-c048-4b5f-b0d5-d018e4af5b1f\") " pod="glance-kuttl-tests/glance-db-create-9t6bx" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.713447 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/a4aaf26d-004b-4fe6-9896-dd0530fdd040-openstack-scripts\") pod \"openstackclient\" (UID: \"a4aaf26d-004b-4fe6-9896-dd0530fdd040\") " pod="glance-kuttl-tests/openstackclient" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.728949 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a4aaf26d-004b-4fe6-9896-dd0530fdd040-openstack-config-secret\") pod \"openstackclient\" (UID: \"a4aaf26d-004b-4fe6-9896-dd0530fdd040\") " pod="glance-kuttl-tests/openstackclient" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.730949 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ck7fh\" (UniqueName: \"kubernetes.io/projected/28c2c617-0d70-4223-a7b3-0043eb3b7d87-kube-api-access-ck7fh\") pod \"glance-f364-account-create-update-xpwg6\" (UID: \"28c2c617-0d70-4223-a7b3-0043eb3b7d87\") " pod="glance-kuttl-tests/glance-f364-account-create-update-xpwg6" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.731880 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mz69s\" (UniqueName: \"kubernetes.io/projected/a4aaf26d-004b-4fe6-9896-dd0530fdd040-kube-api-access-mz69s\") pod \"openstackclient\" (UID: \"a4aaf26d-004b-4fe6-9896-dd0530fdd040\") " pod="glance-kuttl-tests/openstackclient" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.735165 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5q9x\" (UniqueName: \"kubernetes.io/projected/30d08778-c048-4b5f-b0d5-d018e4af5b1f-kube-api-access-d5q9x\") pod \"glance-db-create-9t6bx\" (UID: \"30d08778-c048-4b5f-b0d5-d018e4af5b1f\") " pod="glance-kuttl-tests/glance-db-create-9t6bx" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.851922 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.884523 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-9t6bx" Nov 25 14:04:46 crc kubenswrapper[4702]: I1125 14:04:46.900453 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-f364-account-create-update-xpwg6" Nov 25 14:04:47 crc kubenswrapper[4702]: I1125 14:04:47.342401 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 14:04:47 crc kubenswrapper[4702]: W1125 14:04:47.353753 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4aaf26d_004b_4fe6_9896_dd0530fdd040.slice/crio-bda9d87852d6892e206ea169ab33410dcb4725375bd1ae2cc65e46fc90b73616 WatchSource:0}: Error finding container bda9d87852d6892e206ea169ab33410dcb4725375bd1ae2cc65e46fc90b73616: Status 404 returned error can't find the container with id bda9d87852d6892e206ea169ab33410dcb4725375bd1ae2cc65e46fc90b73616 Nov 25 14:04:47 crc kubenswrapper[4702]: I1125 14:04:47.423916 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-9t6bx"] Nov 25 14:04:47 crc kubenswrapper[4702]: W1125 14:04:47.426543 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30d08778_c048_4b5f_b0d5_d018e4af5b1f.slice/crio-983e3572209504cde1c44a7d31988b501a955f14643ca82610376b4612b825c2 WatchSource:0}: Error finding container 983e3572209504cde1c44a7d31988b501a955f14643ca82610376b4612b825c2: Status 404 returned error can't find the container with id 983e3572209504cde1c44a7d31988b501a955f14643ca82610376b4612b825c2 Nov 25 14:04:47 crc kubenswrapper[4702]: I1125 14:04:47.498850 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-f364-account-create-update-xpwg6"] Nov 25 14:04:47 crc kubenswrapper[4702]: W1125 14:04:47.501472 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28c2c617_0d70_4223_a7b3_0043eb3b7d87.slice/crio-f505bc35eefed4f579da172837fc3a2d303693a623fd41f0cdc6dea3627e5c17 WatchSource:0}: Error finding container f505bc35eefed4f579da172837fc3a2d303693a623fd41f0cdc6dea3627e5c17: Status 404 returned error can't find the container with id f505bc35eefed4f579da172837fc3a2d303693a623fd41f0cdc6dea3627e5c17 Nov 25 14:04:48 crc kubenswrapper[4702]: I1125 14:04:48.007210 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"a4aaf26d-004b-4fe6-9896-dd0530fdd040","Type":"ContainerStarted","Data":"bda9d87852d6892e206ea169ab33410dcb4725375bd1ae2cc65e46fc90b73616"} Nov 25 14:04:48 crc kubenswrapper[4702]: I1125 14:04:48.009521 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-9t6bx" event={"ID":"30d08778-c048-4b5f-b0d5-d018e4af5b1f","Type":"ContainerStarted","Data":"c31eba3d0bcd450abc6b9d57105bb57edb5e7bb354f6e1d085bb6020923360ec"} Nov 25 14:04:48 crc kubenswrapper[4702]: I1125 14:04:48.009567 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-9t6bx" event={"ID":"30d08778-c048-4b5f-b0d5-d018e4af5b1f","Type":"ContainerStarted","Data":"983e3572209504cde1c44a7d31988b501a955f14643ca82610376b4612b825c2"} Nov 25 14:04:48 crc kubenswrapper[4702]: I1125 14:04:48.013349 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-f364-account-create-update-xpwg6" event={"ID":"28c2c617-0d70-4223-a7b3-0043eb3b7d87","Type":"ContainerStarted","Data":"e1700aa0c7a2395cbe33b28a57ba9b6714d59955a966f02ff3b018bdf4db6635"} Nov 25 14:04:48 crc kubenswrapper[4702]: I1125 14:04:48.013412 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-f364-account-create-update-xpwg6" event={"ID":"28c2c617-0d70-4223-a7b3-0043eb3b7d87","Type":"ContainerStarted","Data":"f505bc35eefed4f579da172837fc3a2d303693a623fd41f0cdc6dea3627e5c17"} Nov 25 14:04:48 crc kubenswrapper[4702]: I1125 14:04:48.031683 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-create-9t6bx" podStartSLOduration=2.031664861 podStartE2EDuration="2.031664861s" podCreationTimestamp="2025-11-25 14:04:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 14:04:48.031609579 +0000 UTC m=+1097.194260039" watchObservedRunningTime="2025-11-25 14:04:48.031664861 +0000 UTC m=+1097.194315311" Nov 25 14:04:48 crc kubenswrapper[4702]: I1125 14:04:48.059829 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-f364-account-create-update-xpwg6" podStartSLOduration=2.059798207 podStartE2EDuration="2.059798207s" podCreationTimestamp="2025-11-25 14:04:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 14:04:48.05397165 +0000 UTC m=+1097.216622110" watchObservedRunningTime="2025-11-25 14:04:48.059798207 +0000 UTC m=+1097.222448657" Nov 25 14:04:50 crc kubenswrapper[4702]: I1125 14:04:50.031748 4702 generic.go:334] "Generic (PLEG): container finished" podID="28c2c617-0d70-4223-a7b3-0043eb3b7d87" containerID="e1700aa0c7a2395cbe33b28a57ba9b6714d59955a966f02ff3b018bdf4db6635" exitCode=0 Nov 25 14:04:50 crc kubenswrapper[4702]: I1125 14:04:50.031850 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-f364-account-create-update-xpwg6" event={"ID":"28c2c617-0d70-4223-a7b3-0043eb3b7d87","Type":"ContainerDied","Data":"e1700aa0c7a2395cbe33b28a57ba9b6714d59955a966f02ff3b018bdf4db6635"} Nov 25 14:04:50 crc kubenswrapper[4702]: I1125 14:04:50.034673 4702 generic.go:334] "Generic (PLEG): container finished" podID="30d08778-c048-4b5f-b0d5-d018e4af5b1f" containerID="c31eba3d0bcd450abc6b9d57105bb57edb5e7bb354f6e1d085bb6020923360ec" exitCode=0 Nov 25 14:04:50 crc kubenswrapper[4702]: I1125 14:04:50.034717 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-9t6bx" event={"ID":"30d08778-c048-4b5f-b0d5-d018e4af5b1f","Type":"ContainerDied","Data":"c31eba3d0bcd450abc6b9d57105bb57edb5e7bb354f6e1d085bb6020923360ec"} Nov 25 14:04:56 crc kubenswrapper[4702]: I1125 14:04:56.908311 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-9t6bx" Nov 25 14:04:56 crc kubenswrapper[4702]: I1125 14:04:56.921459 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-f364-account-create-update-xpwg6" Nov 25 14:04:57 crc kubenswrapper[4702]: I1125 14:04:57.033874 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/28c2c617-0d70-4223-a7b3-0043eb3b7d87-operator-scripts\") pod \"28c2c617-0d70-4223-a7b3-0043eb3b7d87\" (UID: \"28c2c617-0d70-4223-a7b3-0043eb3b7d87\") " Nov 25 14:04:57 crc kubenswrapper[4702]: I1125 14:04:57.034712 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28c2c617-0d70-4223-a7b3-0043eb3b7d87-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "28c2c617-0d70-4223-a7b3-0043eb3b7d87" (UID: "28c2c617-0d70-4223-a7b3-0043eb3b7d87"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 14:04:57 crc kubenswrapper[4702]: I1125 14:04:57.034751 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5q9x\" (UniqueName: \"kubernetes.io/projected/30d08778-c048-4b5f-b0d5-d018e4af5b1f-kube-api-access-d5q9x\") pod \"30d08778-c048-4b5f-b0d5-d018e4af5b1f\" (UID: \"30d08778-c048-4b5f-b0d5-d018e4af5b1f\") " Nov 25 14:04:57 crc kubenswrapper[4702]: I1125 14:04:57.034908 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ck7fh\" (UniqueName: \"kubernetes.io/projected/28c2c617-0d70-4223-a7b3-0043eb3b7d87-kube-api-access-ck7fh\") pod \"28c2c617-0d70-4223-a7b3-0043eb3b7d87\" (UID: \"28c2c617-0d70-4223-a7b3-0043eb3b7d87\") " Nov 25 14:04:57 crc kubenswrapper[4702]: I1125 14:04:57.034985 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30d08778-c048-4b5f-b0d5-d018e4af5b1f-operator-scripts\") pod \"30d08778-c048-4b5f-b0d5-d018e4af5b1f\" (UID: \"30d08778-c048-4b5f-b0d5-d018e4af5b1f\") " Nov 25 14:04:57 crc kubenswrapper[4702]: I1125 14:04:57.035253 4702 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/28c2c617-0d70-4223-a7b3-0043eb3b7d87-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 14:04:57 crc kubenswrapper[4702]: I1125 14:04:57.035690 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30d08778-c048-4b5f-b0d5-d018e4af5b1f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "30d08778-c048-4b5f-b0d5-d018e4af5b1f" (UID: "30d08778-c048-4b5f-b0d5-d018e4af5b1f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 14:04:57 crc kubenswrapper[4702]: I1125 14:04:57.038944 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30d08778-c048-4b5f-b0d5-d018e4af5b1f-kube-api-access-d5q9x" (OuterVolumeSpecName: "kube-api-access-d5q9x") pod "30d08778-c048-4b5f-b0d5-d018e4af5b1f" (UID: "30d08778-c048-4b5f-b0d5-d018e4af5b1f"). InnerVolumeSpecName "kube-api-access-d5q9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:04:57 crc kubenswrapper[4702]: I1125 14:04:57.039431 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28c2c617-0d70-4223-a7b3-0043eb3b7d87-kube-api-access-ck7fh" (OuterVolumeSpecName: "kube-api-access-ck7fh") pod "28c2c617-0d70-4223-a7b3-0043eb3b7d87" (UID: "28c2c617-0d70-4223-a7b3-0043eb3b7d87"). InnerVolumeSpecName "kube-api-access-ck7fh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:04:57 crc kubenswrapper[4702]: I1125 14:04:57.102592 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-9t6bx" Nov 25 14:04:57 crc kubenswrapper[4702]: I1125 14:04:57.102604 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-9t6bx" event={"ID":"30d08778-c048-4b5f-b0d5-d018e4af5b1f","Type":"ContainerDied","Data":"983e3572209504cde1c44a7d31988b501a955f14643ca82610376b4612b825c2"} Nov 25 14:04:57 crc kubenswrapper[4702]: I1125 14:04:57.102888 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="983e3572209504cde1c44a7d31988b501a955f14643ca82610376b4612b825c2" Nov 25 14:04:57 crc kubenswrapper[4702]: I1125 14:04:57.105333 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-f364-account-create-update-xpwg6" event={"ID":"28c2c617-0d70-4223-a7b3-0043eb3b7d87","Type":"ContainerDied","Data":"f505bc35eefed4f579da172837fc3a2d303693a623fd41f0cdc6dea3627e5c17"} Nov 25 14:04:57 crc kubenswrapper[4702]: I1125 14:04:57.105372 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f505bc35eefed4f579da172837fc3a2d303693a623fd41f0cdc6dea3627e5c17" Nov 25 14:04:57 crc kubenswrapper[4702]: I1125 14:04:57.105452 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-f364-account-create-update-xpwg6" Nov 25 14:04:57 crc kubenswrapper[4702]: I1125 14:04:57.136092 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5q9x\" (UniqueName: \"kubernetes.io/projected/30d08778-c048-4b5f-b0d5-d018e4af5b1f-kube-api-access-d5q9x\") on node \"crc\" DevicePath \"\"" Nov 25 14:04:57 crc kubenswrapper[4702]: I1125 14:04:57.136137 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ck7fh\" (UniqueName: \"kubernetes.io/projected/28c2c617-0d70-4223-a7b3-0043eb3b7d87-kube-api-access-ck7fh\") on node \"crc\" DevicePath \"\"" Nov 25 14:04:57 crc kubenswrapper[4702]: I1125 14:04:57.136149 4702 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30d08778-c048-4b5f-b0d5-d018e4af5b1f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 14:04:58 crc kubenswrapper[4702]: I1125 14:04:58.116072 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"a4aaf26d-004b-4fe6-9896-dd0530fdd040","Type":"ContainerStarted","Data":"4669cb151a31188c48b753f2fc607c7dc5267c9ce585510213de4beec6d9c4f3"} Nov 25 14:04:58 crc kubenswrapper[4702]: I1125 14:04:58.133599 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstackclient" podStartSLOduration=2.572183882 podStartE2EDuration="12.133578005s" podCreationTimestamp="2025-11-25 14:04:46 +0000 UTC" firstStartedPulling="2025-11-25 14:04:47.358080919 +0000 UTC m=+1096.520731369" lastFinishedPulling="2025-11-25 14:04:56.919475042 +0000 UTC m=+1106.082125492" observedRunningTime="2025-11-25 14:04:58.130680863 +0000 UTC m=+1107.293331333" watchObservedRunningTime="2025-11-25 14:04:58.133578005 +0000 UTC m=+1107.296228455" Nov 25 14:05:01 crc kubenswrapper[4702]: I1125 14:05:01.795439 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-w2jll"] Nov 25 14:05:01 crc kubenswrapper[4702]: E1125 14:05:01.796437 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28c2c617-0d70-4223-a7b3-0043eb3b7d87" containerName="mariadb-account-create-update" Nov 25 14:05:01 crc kubenswrapper[4702]: I1125 14:05:01.796457 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="28c2c617-0d70-4223-a7b3-0043eb3b7d87" containerName="mariadb-account-create-update" Nov 25 14:05:01 crc kubenswrapper[4702]: E1125 14:05:01.796476 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30d08778-c048-4b5f-b0d5-d018e4af5b1f" containerName="mariadb-database-create" Nov 25 14:05:01 crc kubenswrapper[4702]: I1125 14:05:01.796484 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="30d08778-c048-4b5f-b0d5-d018e4af5b1f" containerName="mariadb-database-create" Nov 25 14:05:01 crc kubenswrapper[4702]: I1125 14:05:01.796636 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="28c2c617-0d70-4223-a7b3-0043eb3b7d87" containerName="mariadb-account-create-update" Nov 25 14:05:01 crc kubenswrapper[4702]: I1125 14:05:01.796662 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="30d08778-c048-4b5f-b0d5-d018e4af5b1f" containerName="mariadb-database-create" Nov 25 14:05:01 crc kubenswrapper[4702]: I1125 14:05:01.797363 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-w2jll" Nov 25 14:05:01 crc kubenswrapper[4702]: I1125 14:05:01.799207 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-2f4wc" Nov 25 14:05:01 crc kubenswrapper[4702]: I1125 14:05:01.799669 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 25 14:05:01 crc kubenswrapper[4702]: I1125 14:05:01.809251 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-w2jll"] Nov 25 14:05:01 crc kubenswrapper[4702]: I1125 14:05:01.810417 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15db1634-02df-48f0-a456-a77986ff8e4c-config-data\") pod \"glance-db-sync-w2jll\" (UID: \"15db1634-02df-48f0-a456-a77986ff8e4c\") " pod="glance-kuttl-tests/glance-db-sync-w2jll" Nov 25 14:05:01 crc kubenswrapper[4702]: I1125 14:05:01.810498 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/15db1634-02df-48f0-a456-a77986ff8e4c-db-sync-config-data\") pod \"glance-db-sync-w2jll\" (UID: \"15db1634-02df-48f0-a456-a77986ff8e4c\") " pod="glance-kuttl-tests/glance-db-sync-w2jll" Nov 25 14:05:01 crc kubenswrapper[4702]: I1125 14:05:01.810555 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9d7n2\" (UniqueName: \"kubernetes.io/projected/15db1634-02df-48f0-a456-a77986ff8e4c-kube-api-access-9d7n2\") pod \"glance-db-sync-w2jll\" (UID: \"15db1634-02df-48f0-a456-a77986ff8e4c\") " pod="glance-kuttl-tests/glance-db-sync-w2jll" Nov 25 14:05:01 crc kubenswrapper[4702]: I1125 14:05:01.912178 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15db1634-02df-48f0-a456-a77986ff8e4c-config-data\") pod \"glance-db-sync-w2jll\" (UID: \"15db1634-02df-48f0-a456-a77986ff8e4c\") " pod="glance-kuttl-tests/glance-db-sync-w2jll" Nov 25 14:05:01 crc kubenswrapper[4702]: I1125 14:05:01.912262 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/15db1634-02df-48f0-a456-a77986ff8e4c-db-sync-config-data\") pod \"glance-db-sync-w2jll\" (UID: \"15db1634-02df-48f0-a456-a77986ff8e4c\") " pod="glance-kuttl-tests/glance-db-sync-w2jll" Nov 25 14:05:01 crc kubenswrapper[4702]: I1125 14:05:01.912289 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9d7n2\" (UniqueName: \"kubernetes.io/projected/15db1634-02df-48f0-a456-a77986ff8e4c-kube-api-access-9d7n2\") pod \"glance-db-sync-w2jll\" (UID: \"15db1634-02df-48f0-a456-a77986ff8e4c\") " pod="glance-kuttl-tests/glance-db-sync-w2jll" Nov 25 14:05:01 crc kubenswrapper[4702]: I1125 14:05:01.921092 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15db1634-02df-48f0-a456-a77986ff8e4c-config-data\") pod \"glance-db-sync-w2jll\" (UID: \"15db1634-02df-48f0-a456-a77986ff8e4c\") " pod="glance-kuttl-tests/glance-db-sync-w2jll" Nov 25 14:05:01 crc kubenswrapper[4702]: I1125 14:05:01.923690 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/15db1634-02df-48f0-a456-a77986ff8e4c-db-sync-config-data\") pod \"glance-db-sync-w2jll\" (UID: \"15db1634-02df-48f0-a456-a77986ff8e4c\") " pod="glance-kuttl-tests/glance-db-sync-w2jll" Nov 25 14:05:01 crc kubenswrapper[4702]: I1125 14:05:01.931087 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9d7n2\" (UniqueName: \"kubernetes.io/projected/15db1634-02df-48f0-a456-a77986ff8e4c-kube-api-access-9d7n2\") pod \"glance-db-sync-w2jll\" (UID: \"15db1634-02df-48f0-a456-a77986ff8e4c\") " pod="glance-kuttl-tests/glance-db-sync-w2jll" Nov 25 14:05:02 crc kubenswrapper[4702]: I1125 14:05:02.117834 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-w2jll" Nov 25 14:05:02 crc kubenswrapper[4702]: I1125 14:05:02.622610 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-w2jll"] Nov 25 14:05:03 crc kubenswrapper[4702]: I1125 14:05:03.179631 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-w2jll" event={"ID":"15db1634-02df-48f0-a456-a77986ff8e4c","Type":"ContainerStarted","Data":"0fd607b80e7358752524ca58ecfffd13970ebe03d2c70e56e4c641c816494bd3"} Nov 25 14:05:16 crc kubenswrapper[4702]: I1125 14:05:16.285443 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-w2jll" event={"ID":"15db1634-02df-48f0-a456-a77986ff8e4c","Type":"ContainerStarted","Data":"4d9314c9f8c5153ba1f6da19cd33886fd23f8fdd2b8bfa73337a7a733a20d7d7"} Nov 25 14:05:16 crc kubenswrapper[4702]: I1125 14:05:16.303901 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-w2jll" podStartSLOduration=2.400381139 podStartE2EDuration="15.303883081s" podCreationTimestamp="2025-11-25 14:05:01 +0000 UTC" firstStartedPulling="2025-11-25 14:05:02.640878085 +0000 UTC m=+1111.803528535" lastFinishedPulling="2025-11-25 14:05:15.544380037 +0000 UTC m=+1124.707030477" observedRunningTime="2025-11-25 14:05:16.29861933 +0000 UTC m=+1125.461269780" watchObservedRunningTime="2025-11-25 14:05:16.303883081 +0000 UTC m=+1125.466533531" Nov 25 14:05:57 crc kubenswrapper[4702]: I1125 14:05:57.649702 4702 generic.go:334] "Generic (PLEG): container finished" podID="15db1634-02df-48f0-a456-a77986ff8e4c" containerID="4d9314c9f8c5153ba1f6da19cd33886fd23f8fdd2b8bfa73337a7a733a20d7d7" exitCode=0 Nov 25 14:05:57 crc kubenswrapper[4702]: I1125 14:05:57.649871 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-w2jll" event={"ID":"15db1634-02df-48f0-a456-a77986ff8e4c","Type":"ContainerDied","Data":"4d9314c9f8c5153ba1f6da19cd33886fd23f8fdd2b8bfa73337a7a733a20d7d7"} Nov 25 14:05:58 crc kubenswrapper[4702]: I1125 14:05:58.937167 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-w2jll" Nov 25 14:05:59 crc kubenswrapper[4702]: I1125 14:05:59.013550 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/15db1634-02df-48f0-a456-a77986ff8e4c-db-sync-config-data\") pod \"15db1634-02df-48f0-a456-a77986ff8e4c\" (UID: \"15db1634-02df-48f0-a456-a77986ff8e4c\") " Nov 25 14:05:59 crc kubenswrapper[4702]: I1125 14:05:59.013648 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15db1634-02df-48f0-a456-a77986ff8e4c-config-data\") pod \"15db1634-02df-48f0-a456-a77986ff8e4c\" (UID: \"15db1634-02df-48f0-a456-a77986ff8e4c\") " Nov 25 14:05:59 crc kubenswrapper[4702]: I1125 14:05:59.013694 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9d7n2\" (UniqueName: \"kubernetes.io/projected/15db1634-02df-48f0-a456-a77986ff8e4c-kube-api-access-9d7n2\") pod \"15db1634-02df-48f0-a456-a77986ff8e4c\" (UID: \"15db1634-02df-48f0-a456-a77986ff8e4c\") " Nov 25 14:05:59 crc kubenswrapper[4702]: I1125 14:05:59.020390 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15db1634-02df-48f0-a456-a77986ff8e4c-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "15db1634-02df-48f0-a456-a77986ff8e4c" (UID: "15db1634-02df-48f0-a456-a77986ff8e4c"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:05:59 crc kubenswrapper[4702]: I1125 14:05:59.020427 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15db1634-02df-48f0-a456-a77986ff8e4c-kube-api-access-9d7n2" (OuterVolumeSpecName: "kube-api-access-9d7n2") pod "15db1634-02df-48f0-a456-a77986ff8e4c" (UID: "15db1634-02df-48f0-a456-a77986ff8e4c"). InnerVolumeSpecName "kube-api-access-9d7n2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:05:59 crc kubenswrapper[4702]: I1125 14:05:59.049373 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15db1634-02df-48f0-a456-a77986ff8e4c-config-data" (OuterVolumeSpecName: "config-data") pod "15db1634-02df-48f0-a456-a77986ff8e4c" (UID: "15db1634-02df-48f0-a456-a77986ff8e4c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:05:59 crc kubenswrapper[4702]: I1125 14:05:59.115827 4702 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/15db1634-02df-48f0-a456-a77986ff8e4c-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 14:05:59 crc kubenswrapper[4702]: I1125 14:05:59.115885 4702 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15db1634-02df-48f0-a456-a77986ff8e4c-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 14:05:59 crc kubenswrapper[4702]: I1125 14:05:59.115901 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9d7n2\" (UniqueName: \"kubernetes.io/projected/15db1634-02df-48f0-a456-a77986ff8e4c-kube-api-access-9d7n2\") on node \"crc\" DevicePath \"\"" Nov 25 14:05:59 crc kubenswrapper[4702]: I1125 14:05:59.668915 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-w2jll" event={"ID":"15db1634-02df-48f0-a456-a77986ff8e4c","Type":"ContainerDied","Data":"0fd607b80e7358752524ca58ecfffd13970ebe03d2c70e56e4c641c816494bd3"} Nov 25 14:05:59 crc kubenswrapper[4702]: I1125 14:05:59.668979 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0fd607b80e7358752524ca58ecfffd13970ebe03d2c70e56e4c641c816494bd3" Nov 25 14:05:59 crc kubenswrapper[4702]: I1125 14:05:59.669034 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-w2jll" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.054080 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 14:06:01 crc kubenswrapper[4702]: E1125 14:06:01.055603 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15db1634-02df-48f0-a456-a77986ff8e4c" containerName="glance-db-sync" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.055625 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="15db1634-02df-48f0-a456-a77986ff8e4c" containerName="glance-db-sync" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.055815 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="15db1634-02df-48f0-a456-a77986ff8e4c" containerName="glance-db-sync" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.056959 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.060650 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.060825 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.060880 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-2f4wc" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.077305 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.115878 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.117330 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.142327 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.153603 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.255812 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d968e27b-61a3-4b00-a4d4-79b03bab44ca-logs\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.255874 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-run\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.255899 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-httpd-run\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.255921 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-scripts\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.256041 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-lib-modules\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.256096 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-etc-nvme\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.256154 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.256183 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-config-data\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.256354 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-logs\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.256447 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d968e27b-61a3-4b00-a4d4-79b03bab44ca-config-data\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.256483 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d968e27b-61a3-4b00-a4d4-79b03bab44ca-httpd-run\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.256502 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-dev\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.256577 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.256627 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-run\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.256663 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.256694 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-dev\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.256714 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.256926 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-etc-nvme\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.257089 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.257120 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-sys\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.257164 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-sys\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.257213 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-np7gd\" (UniqueName: \"kubernetes.io/projected/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-kube-api-access-np7gd\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.257289 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.257320 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-lib-modules\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.257345 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d968e27b-61a3-4b00-a4d4-79b03bab44ca-scripts\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.257406 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.257470 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.257563 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbhxt\" (UniqueName: \"kubernetes.io/projected/d968e27b-61a3-4b00-a4d4-79b03bab44ca-kube-api-access-pbhxt\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.258491 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.284768 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.359127 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.359438 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-sys\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.359461 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-sys\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.359482 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-np7gd\" (UniqueName: \"kubernetes.io/projected/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-kube-api-access-np7gd\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.359512 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-lib-modules\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.359532 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d968e27b-61a3-4b00-a4d4-79b03bab44ca-scripts\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.359564 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.359589 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.359620 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbhxt\" (UniqueName: \"kubernetes.io/projected/d968e27b-61a3-4b00-a4d4-79b03bab44ca-kube-api-access-pbhxt\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.359650 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d968e27b-61a3-4b00-a4d4-79b03bab44ca-logs\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.359666 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-run\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.359686 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-httpd-run\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.359702 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-scripts\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.359724 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-lib-modules\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.359778 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-etc-nvme\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.359810 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.359836 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-config-data\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.359860 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-logs\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.359885 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d968e27b-61a3-4b00-a4d4-79b03bab44ca-config-data\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.359911 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d968e27b-61a3-4b00-a4d4-79b03bab44ca-httpd-run\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.359944 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-dev\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.359980 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.360001 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-run\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.360027 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.360048 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-dev\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.360066 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.360089 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-etc-nvme\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.360279 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-etc-nvme\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.359377 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.361085 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-lib-modules\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.361152 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-sys\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.361179 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-sys\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.361591 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-lib-modules\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.365830 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d968e27b-61a3-4b00-a4d4-79b03bab44ca-httpd-run\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.366192 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.366255 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-etc-nvme\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.366299 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") device mount path \"/mnt/openstack/pv02\"" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.366362 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.366390 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d968e27b-61a3-4b00-a4d4-79b03bab44ca-logs\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.366955 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.367009 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-run\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.367043 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-dev\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.367185 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-logs\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.367475 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-httpd-run\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.367522 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-run\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.367573 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-dev\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.367617 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.369971 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.372445 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-scripts\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.372806 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d968e27b-61a3-4b00-a4d4-79b03bab44ca-scripts\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.373005 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-config-data\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.378467 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d968e27b-61a3-4b00-a4d4-79b03bab44ca-config-data\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.381602 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-np7gd\" (UniqueName: \"kubernetes.io/projected/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-kube-api-access-np7gd\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.390353 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.396613 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbhxt\" (UniqueName: \"kubernetes.io/projected/d968e27b-61a3-4b00-a4d4-79b03bab44ca-kube-api-access-pbhxt\") pod \"glance-default-single-0\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.399976 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.405333 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-1\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.433535 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.679158 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:01 crc kubenswrapper[4702]: I1125 14:06:01.959769 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 14:06:02 crc kubenswrapper[4702]: I1125 14:06:02.189394 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 14:06:02 crc kubenswrapper[4702]: W1125 14:06:02.194474 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcbb1bc07_9697_4f91_bfa2_b00fcb85fe5e.slice/crio-ec2c5568defe7892250e025ca66b0f2f9cedc9130385a4539250095907535e42 WatchSource:0}: Error finding container ec2c5568defe7892250e025ca66b0f2f9cedc9130385a4539250095907535e42: Status 404 returned error can't find the container with id ec2c5568defe7892250e025ca66b0f2f9cedc9130385a4539250095907535e42 Nov 25 14:06:02 crc kubenswrapper[4702]: I1125 14:06:02.696302 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"d968e27b-61a3-4b00-a4d4-79b03bab44ca","Type":"ContainerStarted","Data":"39280e444daf5800e562039ce960a07e28d2db7d9c861efebea90229557dc19f"} Nov 25 14:06:02 crc kubenswrapper[4702]: I1125 14:06:02.696714 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"d968e27b-61a3-4b00-a4d4-79b03bab44ca","Type":"ContainerStarted","Data":"5e5aeb78fcb471c169f1664c3c88adde7c582d289e29137cce2b3e86d7e4e004"} Nov 25 14:06:02 crc kubenswrapper[4702]: I1125 14:06:02.696726 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"d968e27b-61a3-4b00-a4d4-79b03bab44ca","Type":"ContainerStarted","Data":"8e887a52bf48665b2cd572dcbfe9eb84a4744bcfd2d340a2f1905afb698e0b24"} Nov 25 14:06:02 crc kubenswrapper[4702]: I1125 14:06:02.698401 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e","Type":"ContainerStarted","Data":"e8a739726777867d70cfc1d62be96bb994733c4ff18896df97aedbcbaa3b799a"} Nov 25 14:06:02 crc kubenswrapper[4702]: I1125 14:06:02.698549 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e","Type":"ContainerStarted","Data":"41f52290460e870d5a774f97567d5922d97d5a39d98b2a7853314fb70f9bf11d"} Nov 25 14:06:02 crc kubenswrapper[4702]: I1125 14:06:02.698907 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e","Type":"ContainerStarted","Data":"ec2c5568defe7892250e025ca66b0f2f9cedc9130385a4539250095907535e42"} Nov 25 14:06:02 crc kubenswrapper[4702]: I1125 14:06:02.722989 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=2.72295841 podStartE2EDuration="2.72295841s" podCreationTimestamp="2025-11-25 14:06:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 14:06:02.720611572 +0000 UTC m=+1171.883262022" watchObservedRunningTime="2025-11-25 14:06:02.72295841 +0000 UTC m=+1171.885608860" Nov 25 14:06:02 crc kubenswrapper[4702]: I1125 14:06:02.772185 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-1" podStartSLOduration=1.7721538749999999 podStartE2EDuration="1.772153875s" podCreationTimestamp="2025-11-25 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 14:06:02.742552052 +0000 UTC m=+1171.905202522" watchObservedRunningTime="2025-11-25 14:06:02.772153875 +0000 UTC m=+1171.934804325" Nov 25 14:06:11 crc kubenswrapper[4702]: I1125 14:06:11.434333 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:11 crc kubenswrapper[4702]: I1125 14:06:11.435045 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:11 crc kubenswrapper[4702]: I1125 14:06:11.464617 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:11 crc kubenswrapper[4702]: I1125 14:06:11.478796 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:11 crc kubenswrapper[4702]: I1125 14:06:11.680635 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:11 crc kubenswrapper[4702]: I1125 14:06:11.680703 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:11 crc kubenswrapper[4702]: I1125 14:06:11.716150 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:11 crc kubenswrapper[4702]: I1125 14:06:11.721197 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:11 crc kubenswrapper[4702]: I1125 14:06:11.771394 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:11 crc kubenswrapper[4702]: I1125 14:06:11.771612 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:11 crc kubenswrapper[4702]: I1125 14:06:11.771657 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:11 crc kubenswrapper[4702]: I1125 14:06:11.771667 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:13 crc kubenswrapper[4702]: I1125 14:06:13.881742 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:13 crc kubenswrapper[4702]: I1125 14:06:13.882335 4702 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 14:06:13 crc kubenswrapper[4702]: I1125 14:06:13.905247 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:06:13 crc kubenswrapper[4702]: I1125 14:06:13.959134 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 14:06:13 crc kubenswrapper[4702]: I1125 14:06:13.959323 4702 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 14:06:13 crc kubenswrapper[4702]: I1125 14:06:13.959337 4702 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 14:06:13 crc kubenswrapper[4702]: I1125 14:06:13.959881 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="d968e27b-61a3-4b00-a4d4-79b03bab44ca" containerName="glance-httpd" containerID="cri-o://39280e444daf5800e562039ce960a07e28d2db7d9c861efebea90229557dc19f" gracePeriod=30 Nov 25 14:06:13 crc kubenswrapper[4702]: I1125 14:06:13.959878 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="d968e27b-61a3-4b00-a4d4-79b03bab44ca" containerName="glance-log" containerID="cri-o://5e5aeb78fcb471c169f1664c3c88adde7c582d289e29137cce2b3e86d7e4e004" gracePeriod=30 Nov 25 14:06:13 crc kubenswrapper[4702]: I1125 14:06:13.984444 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="d968e27b-61a3-4b00-a4d4-79b03bab44ca" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.0.101:9292/healthcheck\": EOF" Nov 25 14:06:14 crc kubenswrapper[4702]: I1125 14:06:14.189672 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="d968e27b-61a3-4b00-a4d4-79b03bab44ca" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.0.101:9292/healthcheck\": EOF" Nov 25 14:06:14 crc kubenswrapper[4702]: I1125 14:06:14.798680 4702 generic.go:334] "Generic (PLEG): container finished" podID="d968e27b-61a3-4b00-a4d4-79b03bab44ca" containerID="5e5aeb78fcb471c169f1664c3c88adde7c582d289e29137cce2b3e86d7e4e004" exitCode=143 Nov 25 14:06:14 crc kubenswrapper[4702]: I1125 14:06:14.798793 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"d968e27b-61a3-4b00-a4d4-79b03bab44ca","Type":"ContainerDied","Data":"5e5aeb78fcb471c169f1664c3c88adde7c582d289e29137cce2b3e86d7e4e004"} Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.366284 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="d968e27b-61a3-4b00-a4d4-79b03bab44ca" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.0.101:9292/healthcheck\": read tcp 10.217.0.2:46918->10.217.0.101:9292: read: connection reset by peer" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.366342 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="d968e27b-61a3-4b00-a4d4-79b03bab44ca" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.0.101:9292/healthcheck\": read tcp 10.217.0.2:46920->10.217.0.101:9292: read: connection reset by peer" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.701753 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.823594 4702 generic.go:334] "Generic (PLEG): container finished" podID="d968e27b-61a3-4b00-a4d4-79b03bab44ca" containerID="39280e444daf5800e562039ce960a07e28d2db7d9c861efebea90229557dc19f" exitCode=0 Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.823660 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.823657 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"d968e27b-61a3-4b00-a4d4-79b03bab44ca","Type":"ContainerDied","Data":"39280e444daf5800e562039ce960a07e28d2db7d9c861efebea90229557dc19f"} Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.823858 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"d968e27b-61a3-4b00-a4d4-79b03bab44ca","Type":"ContainerDied","Data":"8e887a52bf48665b2cd572dcbfe9eb84a4744bcfd2d340a2f1905afb698e0b24"} Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.823918 4702 scope.go:117] "RemoveContainer" containerID="39280e444daf5800e562039ce960a07e28d2db7d9c861efebea90229557dc19f" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.853191 4702 scope.go:117] "RemoveContainer" containerID="5e5aeb78fcb471c169f1664c3c88adde7c582d289e29137cce2b3e86d7e4e004" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.883836 4702 scope.go:117] "RemoveContainer" containerID="39280e444daf5800e562039ce960a07e28d2db7d9c861efebea90229557dc19f" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.884293 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-lib-modules\") pod \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.884551 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-run\") pod \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.884594 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d968e27b-61a3-4b00-a4d4-79b03bab44ca-logs\") pod \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.884677 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.884706 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-sys\") pod \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.884751 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d968e27b-61a3-4b00-a4d4-79b03bab44ca-config-data\") pod \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.884776 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d968e27b-61a3-4b00-a4d4-79b03bab44ca-httpd-run\") pod \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.884805 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-dev\") pod \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.884831 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbhxt\" (UniqueName: \"kubernetes.io/projected/d968e27b-61a3-4b00-a4d4-79b03bab44ca-kube-api-access-pbhxt\") pod \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.884858 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-var-locks-brick\") pod \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.884887 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-etc-iscsi\") pod \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.884916 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-sys" (OuterVolumeSpecName: "sys") pod "d968e27b-61a3-4b00-a4d4-79b03bab44ca" (UID: "d968e27b-61a3-4b00-a4d4-79b03bab44ca"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.884945 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d968e27b-61a3-4b00-a4d4-79b03bab44ca-scripts\") pod \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.884969 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "d968e27b-61a3-4b00-a4d4-79b03bab44ca" (UID: "d968e27b-61a3-4b00-a4d4-79b03bab44ca"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.884985 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.885004 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-run" (OuterVolumeSpecName: "run") pod "d968e27b-61a3-4b00-a4d4-79b03bab44ca" (UID: "d968e27b-61a3-4b00-a4d4-79b03bab44ca"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.885019 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-etc-nvme\") pod \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\" (UID: \"d968e27b-61a3-4b00-a4d4-79b03bab44ca\") " Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.885404 4702 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-run\") on node \"crc\" DevicePath \"\"" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.885420 4702 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-sys\") on node \"crc\" DevicePath \"\"" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.885433 4702 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.885452 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d968e27b-61a3-4b00-a4d4-79b03bab44ca-logs" (OuterVolumeSpecName: "logs") pod "d968e27b-61a3-4b00-a4d4-79b03bab44ca" (UID: "d968e27b-61a3-4b00-a4d4-79b03bab44ca"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.885499 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "d968e27b-61a3-4b00-a4d4-79b03bab44ca" (UID: "d968e27b-61a3-4b00-a4d4-79b03bab44ca"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:06:17 crc kubenswrapper[4702]: E1125 14:06:17.884827 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39280e444daf5800e562039ce960a07e28d2db7d9c861efebea90229557dc19f\": container with ID starting with 39280e444daf5800e562039ce960a07e28d2db7d9c861efebea90229557dc19f not found: ID does not exist" containerID="39280e444daf5800e562039ce960a07e28d2db7d9c861efebea90229557dc19f" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.886776 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39280e444daf5800e562039ce960a07e28d2db7d9c861efebea90229557dc19f"} err="failed to get container status \"39280e444daf5800e562039ce960a07e28d2db7d9c861efebea90229557dc19f\": rpc error: code = NotFound desc = could not find container \"39280e444daf5800e562039ce960a07e28d2db7d9c861efebea90229557dc19f\": container with ID starting with 39280e444daf5800e562039ce960a07e28d2db7d9c861efebea90229557dc19f not found: ID does not exist" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.886825 4702 scope.go:117] "RemoveContainer" containerID="5e5aeb78fcb471c169f1664c3c88adde7c582d289e29137cce2b3e86d7e4e004" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.887372 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d968e27b-61a3-4b00-a4d4-79b03bab44ca-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d968e27b-61a3-4b00-a4d4-79b03bab44ca" (UID: "d968e27b-61a3-4b00-a4d4-79b03bab44ca"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.887358 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "d968e27b-61a3-4b00-a4d4-79b03bab44ca" (UID: "d968e27b-61a3-4b00-a4d4-79b03bab44ca"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.887420 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-dev" (OuterVolumeSpecName: "dev") pod "d968e27b-61a3-4b00-a4d4-79b03bab44ca" (UID: "d968e27b-61a3-4b00-a4d4-79b03bab44ca"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.887457 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "d968e27b-61a3-4b00-a4d4-79b03bab44ca" (UID: "d968e27b-61a3-4b00-a4d4-79b03bab44ca"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:06:17 crc kubenswrapper[4702]: E1125 14:06:17.891666 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e5aeb78fcb471c169f1664c3c88adde7c582d289e29137cce2b3e86d7e4e004\": container with ID starting with 5e5aeb78fcb471c169f1664c3c88adde7c582d289e29137cce2b3e86d7e4e004 not found: ID does not exist" containerID="5e5aeb78fcb471c169f1664c3c88adde7c582d289e29137cce2b3e86d7e4e004" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.891926 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e5aeb78fcb471c169f1664c3c88adde7c582d289e29137cce2b3e86d7e4e004"} err="failed to get container status \"5e5aeb78fcb471c169f1664c3c88adde7c582d289e29137cce2b3e86d7e4e004\": rpc error: code = NotFound desc = could not find container \"5e5aeb78fcb471c169f1664c3c88adde7c582d289e29137cce2b3e86d7e4e004\": container with ID starting with 5e5aeb78fcb471c169f1664c3c88adde7c582d289e29137cce2b3e86d7e4e004 not found: ID does not exist" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.895257 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "d968e27b-61a3-4b00-a4d4-79b03bab44ca" (UID: "d968e27b-61a3-4b00-a4d4-79b03bab44ca"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.895321 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance-cache") pod "d968e27b-61a3-4b00-a4d4-79b03bab44ca" (UID: "d968e27b-61a3-4b00-a4d4-79b03bab44ca"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.895434 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d968e27b-61a3-4b00-a4d4-79b03bab44ca-scripts" (OuterVolumeSpecName: "scripts") pod "d968e27b-61a3-4b00-a4d4-79b03bab44ca" (UID: "d968e27b-61a3-4b00-a4d4-79b03bab44ca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.895437 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d968e27b-61a3-4b00-a4d4-79b03bab44ca-kube-api-access-pbhxt" (OuterVolumeSpecName: "kube-api-access-pbhxt") pod "d968e27b-61a3-4b00-a4d4-79b03bab44ca" (UID: "d968e27b-61a3-4b00-a4d4-79b03bab44ca"). InnerVolumeSpecName "kube-api-access-pbhxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.928337 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d968e27b-61a3-4b00-a4d4-79b03bab44ca-config-data" (OuterVolumeSpecName: "config-data") pod "d968e27b-61a3-4b00-a4d4-79b03bab44ca" (UID: "d968e27b-61a3-4b00-a4d4-79b03bab44ca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.987087 4702 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d968e27b-61a3-4b00-a4d4-79b03bab44ca-logs\") on node \"crc\" DevicePath \"\"" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.987156 4702 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.987175 4702 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d968e27b-61a3-4b00-a4d4-79b03bab44ca-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.987187 4702 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d968e27b-61a3-4b00-a4d4-79b03bab44ca-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.987274 4702 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-dev\") on node \"crc\" DevicePath \"\"" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.987288 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbhxt\" (UniqueName: \"kubernetes.io/projected/d968e27b-61a3-4b00-a4d4-79b03bab44ca-kube-api-access-pbhxt\") on node \"crc\" DevicePath \"\"" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.987302 4702 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.987313 4702 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.987324 4702 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d968e27b-61a3-4b00-a4d4-79b03bab44ca-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.987343 4702 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 25 14:06:17 crc kubenswrapper[4702]: I1125 14:06:17.987355 4702 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d968e27b-61a3-4b00-a4d4-79b03bab44ca-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.001247 4702 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.002387 4702 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.089779 4702 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.089813 4702 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.155697 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.164023 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.180297 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 14:06:18 crc kubenswrapper[4702]: E1125 14:06:18.180617 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d968e27b-61a3-4b00-a4d4-79b03bab44ca" containerName="glance-httpd" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.180633 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="d968e27b-61a3-4b00-a4d4-79b03bab44ca" containerName="glance-httpd" Nov 25 14:06:18 crc kubenswrapper[4702]: E1125 14:06:18.180642 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d968e27b-61a3-4b00-a4d4-79b03bab44ca" containerName="glance-log" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.180649 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="d968e27b-61a3-4b00-a4d4-79b03bab44ca" containerName="glance-log" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.180797 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="d968e27b-61a3-4b00-a4d4-79b03bab44ca" containerName="glance-httpd" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.180814 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="d968e27b-61a3-4b00-a4d4-79b03bab44ca" containerName="glance-log" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.181627 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.213502 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 14:06:18 crc kubenswrapper[4702]: E1125 14:06:18.256553 4702 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd968e27b_61a3_4b00_a4d4_79b03bab44ca.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd968e27b_61a3_4b00_a4d4_79b03bab44ca.slice/crio-8e887a52bf48665b2cd572dcbfe9eb84a4744bcfd2d340a2f1905afb698e0b24\": RecentStats: unable to find data in memory cache]" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.293147 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-dev\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.293566 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62nwp\" (UniqueName: \"kubernetes.io/projected/3813201d-36d3-4599-9d57-d5b8a5b96f13-kube-api-access-62nwp\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.293591 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-sys\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.293629 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.293643 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-lib-modules\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.293667 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3813201d-36d3-4599-9d57-d5b8a5b96f13-httpd-run\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.293693 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.293941 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-etc-nvme\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.294049 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3813201d-36d3-4599-9d57-d5b8a5b96f13-logs\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.294126 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3813201d-36d3-4599-9d57-d5b8a5b96f13-config-data\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.294182 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-run\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.294280 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.294323 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3813201d-36d3-4599-9d57-d5b8a5b96f13-scripts\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.294492 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.396335 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-dev\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.396396 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62nwp\" (UniqueName: \"kubernetes.io/projected/3813201d-36d3-4599-9d57-d5b8a5b96f13-kube-api-access-62nwp\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.396419 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-sys\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.396459 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-lib-modules\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.396460 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-dev\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.396479 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.396529 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3813201d-36d3-4599-9d57-d5b8a5b96f13-httpd-run\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.396557 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.396583 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-etc-nvme\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.396607 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3813201d-36d3-4599-9d57-d5b8a5b96f13-logs\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.396632 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3813201d-36d3-4599-9d57-d5b8a5b96f13-config-data\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.396654 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-run\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.396683 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.396706 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3813201d-36d3-4599-9d57-d5b8a5b96f13-scripts\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.396734 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.396756 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.396838 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.396941 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.397014 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-run\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.397054 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-sys\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.397389 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3813201d-36d3-4599-9d57-d5b8a5b96f13-logs\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.397659 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3813201d-36d3-4599-9d57-d5b8a5b96f13-httpd-run\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.397777 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.399199 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-etc-nvme\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.399289 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-lib-modules\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.406677 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3813201d-36d3-4599-9d57-d5b8a5b96f13-config-data\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.412200 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3813201d-36d3-4599-9d57-d5b8a5b96f13-scripts\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.416609 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62nwp\" (UniqueName: \"kubernetes.io/projected/3813201d-36d3-4599-9d57-d5b8a5b96f13-kube-api-access-62nwp\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.421306 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.425517 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-0\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.500908 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:18 crc kubenswrapper[4702]: I1125 14:06:18.930734 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 14:06:18 crc kubenswrapper[4702]: W1125 14:06:18.945275 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3813201d_36d3_4599_9d57_d5b8a5b96f13.slice/crio-7b8d7f358de9dae84426267a1aa30c8a9b290374a205102d58df06a44735fcc2 WatchSource:0}: Error finding container 7b8d7f358de9dae84426267a1aa30c8a9b290374a205102d58df06a44735fcc2: Status 404 returned error can't find the container with id 7b8d7f358de9dae84426267a1aa30c8a9b290374a205102d58df06a44735fcc2 Nov 25 14:06:19 crc kubenswrapper[4702]: I1125 14:06:19.814051 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d968e27b-61a3-4b00-a4d4-79b03bab44ca" path="/var/lib/kubelet/pods/d968e27b-61a3-4b00-a4d4-79b03bab44ca/volumes" Nov 25 14:06:19 crc kubenswrapper[4702]: I1125 14:06:19.844758 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"3813201d-36d3-4599-9d57-d5b8a5b96f13","Type":"ContainerStarted","Data":"ea926cd6e89698d0a6c7550aeffd0d571d9babf3ba5b35e0320cfe15a608addd"} Nov 25 14:06:19 crc kubenswrapper[4702]: I1125 14:06:19.844807 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"3813201d-36d3-4599-9d57-d5b8a5b96f13","Type":"ContainerStarted","Data":"80e2dd0a5e92cadc9c3dfad4aa2e542f1c5dfa142baea0466c9c4ccb1c41b76a"} Nov 25 14:06:19 crc kubenswrapper[4702]: I1125 14:06:19.844817 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"3813201d-36d3-4599-9d57-d5b8a5b96f13","Type":"ContainerStarted","Data":"7b8d7f358de9dae84426267a1aa30c8a9b290374a205102d58df06a44735fcc2"} Nov 25 14:06:19 crc kubenswrapper[4702]: I1125 14:06:19.879915 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=1.879894159 podStartE2EDuration="1.879894159s" podCreationTimestamp="2025-11-25 14:06:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 14:06:19.877657624 +0000 UTC m=+1189.040308084" watchObservedRunningTime="2025-11-25 14:06:19.879894159 +0000 UTC m=+1189.042544599" Nov 25 14:06:28 crc kubenswrapper[4702]: I1125 14:06:28.501779 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:28 crc kubenswrapper[4702]: I1125 14:06:28.502653 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:28 crc kubenswrapper[4702]: I1125 14:06:28.530355 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:28 crc kubenswrapper[4702]: I1125 14:06:28.543695 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:28 crc kubenswrapper[4702]: I1125 14:06:28.919088 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:28 crc kubenswrapper[4702]: I1125 14:06:28.919174 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:31 crc kubenswrapper[4702]: I1125 14:06:31.148447 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:31 crc kubenswrapper[4702]: I1125 14:06:31.149323 4702 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 14:06:31 crc kubenswrapper[4702]: I1125 14:06:31.250856 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:06:33 crc kubenswrapper[4702]: I1125 14:06:33.963448 4702 patch_prober.go:28] interesting pod/machine-config-daemon-qqr76 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 14:06:33 crc kubenswrapper[4702]: I1125 14:06:33.964310 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 14:07:01 crc kubenswrapper[4702]: I1125 14:07:01.362666 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-w2jll"] Nov 25 14:07:01 crc kubenswrapper[4702]: I1125 14:07:01.368305 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-w2jll"] Nov 25 14:07:01 crc kubenswrapper[4702]: I1125 14:07:01.411750 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glancef364-account-delete-26dg2"] Nov 25 14:07:01 crc kubenswrapper[4702]: I1125 14:07:01.412683 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancef364-account-delete-26dg2" Nov 25 14:07:01 crc kubenswrapper[4702]: I1125 14:07:01.425809 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glancef364-account-delete-26dg2"] Nov 25 14:07:01 crc kubenswrapper[4702]: I1125 14:07:01.496915 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 14:07:01 crc kubenswrapper[4702]: I1125 14:07:01.497351 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="3813201d-36d3-4599-9d57-d5b8a5b96f13" containerName="glance-log" containerID="cri-o://80e2dd0a5e92cadc9c3dfad4aa2e542f1c5dfa142baea0466c9c4ccb1c41b76a" gracePeriod=30 Nov 25 14:07:01 crc kubenswrapper[4702]: I1125 14:07:01.497880 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="3813201d-36d3-4599-9d57-d5b8a5b96f13" containerName="glance-httpd" containerID="cri-o://ea926cd6e89698d0a6c7550aeffd0d571d9babf3ba5b35e0320cfe15a608addd" gracePeriod=30 Nov 25 14:07:01 crc kubenswrapper[4702]: I1125 14:07:01.498516 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32b081d0-f397-4b90-bae2-8a398aa32084-operator-scripts\") pod \"glancef364-account-delete-26dg2\" (UID: \"32b081d0-f397-4b90-bae2-8a398aa32084\") " pod="glance-kuttl-tests/glancef364-account-delete-26dg2" Nov 25 14:07:01 crc kubenswrapper[4702]: I1125 14:07:01.498711 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bv42\" (UniqueName: \"kubernetes.io/projected/32b081d0-f397-4b90-bae2-8a398aa32084-kube-api-access-8bv42\") pod \"glancef364-account-delete-26dg2\" (UID: \"32b081d0-f397-4b90-bae2-8a398aa32084\") " pod="glance-kuttl-tests/glancef364-account-delete-26dg2" Nov 25 14:07:01 crc kubenswrapper[4702]: I1125 14:07:01.509585 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 14:07:01 crc kubenswrapper[4702]: I1125 14:07:01.509861 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e" containerName="glance-log" containerID="cri-o://41f52290460e870d5a774f97567d5922d97d5a39d98b2a7853314fb70f9bf11d" gracePeriod=30 Nov 25 14:07:01 crc kubenswrapper[4702]: I1125 14:07:01.510297 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e" containerName="glance-httpd" containerID="cri-o://e8a739726777867d70cfc1d62be96bb994733c4ff18896df97aedbcbaa3b799a" gracePeriod=30 Nov 25 14:07:01 crc kubenswrapper[4702]: I1125 14:07:01.595156 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 14:07:01 crc kubenswrapper[4702]: I1125 14:07:01.595864 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/openstackclient" podUID="a4aaf26d-004b-4fe6-9896-dd0530fdd040" containerName="openstackclient" containerID="cri-o://4669cb151a31188c48b753f2fc607c7dc5267c9ce585510213de4beec6d9c4f3" gracePeriod=30 Nov 25 14:07:01 crc kubenswrapper[4702]: I1125 14:07:01.600328 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bv42\" (UniqueName: \"kubernetes.io/projected/32b081d0-f397-4b90-bae2-8a398aa32084-kube-api-access-8bv42\") pod \"glancef364-account-delete-26dg2\" (UID: \"32b081d0-f397-4b90-bae2-8a398aa32084\") " pod="glance-kuttl-tests/glancef364-account-delete-26dg2" Nov 25 14:07:01 crc kubenswrapper[4702]: I1125 14:07:01.600450 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32b081d0-f397-4b90-bae2-8a398aa32084-operator-scripts\") pod \"glancef364-account-delete-26dg2\" (UID: \"32b081d0-f397-4b90-bae2-8a398aa32084\") " pod="glance-kuttl-tests/glancef364-account-delete-26dg2" Nov 25 14:07:01 crc kubenswrapper[4702]: I1125 14:07:01.601389 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32b081d0-f397-4b90-bae2-8a398aa32084-operator-scripts\") pod \"glancef364-account-delete-26dg2\" (UID: \"32b081d0-f397-4b90-bae2-8a398aa32084\") " pod="glance-kuttl-tests/glancef364-account-delete-26dg2" Nov 25 14:07:01 crc kubenswrapper[4702]: I1125 14:07:01.633291 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bv42\" (UniqueName: \"kubernetes.io/projected/32b081d0-f397-4b90-bae2-8a398aa32084-kube-api-access-8bv42\") pod \"glancef364-account-delete-26dg2\" (UID: \"32b081d0-f397-4b90-bae2-8a398aa32084\") " pod="glance-kuttl-tests/glancef364-account-delete-26dg2" Nov 25 14:07:01 crc kubenswrapper[4702]: I1125 14:07:01.735593 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancef364-account-delete-26dg2" Nov 25 14:07:01 crc kubenswrapper[4702]: I1125 14:07:01.818862 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15db1634-02df-48f0-a456-a77986ff8e4c" path="/var/lib/kubelet/pods/15db1634-02df-48f0-a456-a77986ff8e4c/volumes" Nov 25 14:07:02 crc kubenswrapper[4702]: I1125 14:07:02.102253 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 25 14:07:02 crc kubenswrapper[4702]: I1125 14:07:02.196050 4702 generic.go:334] "Generic (PLEG): container finished" podID="a4aaf26d-004b-4fe6-9896-dd0530fdd040" containerID="4669cb151a31188c48b753f2fc607c7dc5267c9ce585510213de4beec6d9c4f3" exitCode=143 Nov 25 14:07:02 crc kubenswrapper[4702]: I1125 14:07:02.196128 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 25 14:07:02 crc kubenswrapper[4702]: I1125 14:07:02.196130 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"a4aaf26d-004b-4fe6-9896-dd0530fdd040","Type":"ContainerDied","Data":"4669cb151a31188c48b753f2fc607c7dc5267c9ce585510213de4beec6d9c4f3"} Nov 25 14:07:02 crc kubenswrapper[4702]: I1125 14:07:02.196287 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"a4aaf26d-004b-4fe6-9896-dd0530fdd040","Type":"ContainerDied","Data":"bda9d87852d6892e206ea169ab33410dcb4725375bd1ae2cc65e46fc90b73616"} Nov 25 14:07:02 crc kubenswrapper[4702]: I1125 14:07:02.196346 4702 scope.go:117] "RemoveContainer" containerID="4669cb151a31188c48b753f2fc607c7dc5267c9ce585510213de4beec6d9c4f3" Nov 25 14:07:02 crc kubenswrapper[4702]: I1125 14:07:02.198552 4702 generic.go:334] "Generic (PLEG): container finished" podID="3813201d-36d3-4599-9d57-d5b8a5b96f13" containerID="80e2dd0a5e92cadc9c3dfad4aa2e542f1c5dfa142baea0466c9c4ccb1c41b76a" exitCode=143 Nov 25 14:07:02 crc kubenswrapper[4702]: I1125 14:07:02.198592 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"3813201d-36d3-4599-9d57-d5b8a5b96f13","Type":"ContainerDied","Data":"80e2dd0a5e92cadc9c3dfad4aa2e542f1c5dfa142baea0466c9c4ccb1c41b76a"} Nov 25 14:07:02 crc kubenswrapper[4702]: I1125 14:07:02.200530 4702 generic.go:334] "Generic (PLEG): container finished" podID="cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e" containerID="41f52290460e870d5a774f97567d5922d97d5a39d98b2a7853314fb70f9bf11d" exitCode=143 Nov 25 14:07:02 crc kubenswrapper[4702]: I1125 14:07:02.200581 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e","Type":"ContainerDied","Data":"41f52290460e870d5a774f97567d5922d97d5a39d98b2a7853314fb70f9bf11d"} Nov 25 14:07:02 crc kubenswrapper[4702]: I1125 14:07:02.217041 4702 scope.go:117] "RemoveContainer" containerID="4669cb151a31188c48b753f2fc607c7dc5267c9ce585510213de4beec6d9c4f3" Nov 25 14:07:02 crc kubenswrapper[4702]: I1125 14:07:02.217444 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a4aaf26d-004b-4fe6-9896-dd0530fdd040-openstack-config-secret\") pod \"a4aaf26d-004b-4fe6-9896-dd0530fdd040\" (UID: \"a4aaf26d-004b-4fe6-9896-dd0530fdd040\") " Nov 25 14:07:02 crc kubenswrapper[4702]: I1125 14:07:02.217558 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/a4aaf26d-004b-4fe6-9896-dd0530fdd040-openstack-scripts\") pod \"a4aaf26d-004b-4fe6-9896-dd0530fdd040\" (UID: \"a4aaf26d-004b-4fe6-9896-dd0530fdd040\") " Nov 25 14:07:02 crc kubenswrapper[4702]: I1125 14:07:02.217621 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a4aaf26d-004b-4fe6-9896-dd0530fdd040-openstack-config\") pod \"a4aaf26d-004b-4fe6-9896-dd0530fdd040\" (UID: \"a4aaf26d-004b-4fe6-9896-dd0530fdd040\") " Nov 25 14:07:02 crc kubenswrapper[4702]: E1125 14:07:02.217658 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4669cb151a31188c48b753f2fc607c7dc5267c9ce585510213de4beec6d9c4f3\": container with ID starting with 4669cb151a31188c48b753f2fc607c7dc5267c9ce585510213de4beec6d9c4f3 not found: ID does not exist" containerID="4669cb151a31188c48b753f2fc607c7dc5267c9ce585510213de4beec6d9c4f3" Nov 25 14:07:02 crc kubenswrapper[4702]: I1125 14:07:02.217691 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4669cb151a31188c48b753f2fc607c7dc5267c9ce585510213de4beec6d9c4f3"} err="failed to get container status \"4669cb151a31188c48b753f2fc607c7dc5267c9ce585510213de4beec6d9c4f3\": rpc error: code = NotFound desc = could not find container \"4669cb151a31188c48b753f2fc607c7dc5267c9ce585510213de4beec6d9c4f3\": container with ID starting with 4669cb151a31188c48b753f2fc607c7dc5267c9ce585510213de4beec6d9c4f3 not found: ID does not exist" Nov 25 14:07:02 crc kubenswrapper[4702]: I1125 14:07:02.217734 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mz69s\" (UniqueName: \"kubernetes.io/projected/a4aaf26d-004b-4fe6-9896-dd0530fdd040-kube-api-access-mz69s\") pod \"a4aaf26d-004b-4fe6-9896-dd0530fdd040\" (UID: \"a4aaf26d-004b-4fe6-9896-dd0530fdd040\") " Nov 25 14:07:02 crc kubenswrapper[4702]: I1125 14:07:02.218703 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4aaf26d-004b-4fe6-9896-dd0530fdd040-openstack-scripts" (OuterVolumeSpecName: "openstack-scripts") pod "a4aaf26d-004b-4fe6-9896-dd0530fdd040" (UID: "a4aaf26d-004b-4fe6-9896-dd0530fdd040"). InnerVolumeSpecName "openstack-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 14:07:02 crc kubenswrapper[4702]: I1125 14:07:02.224189 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4aaf26d-004b-4fe6-9896-dd0530fdd040-kube-api-access-mz69s" (OuterVolumeSpecName: "kube-api-access-mz69s") pod "a4aaf26d-004b-4fe6-9896-dd0530fdd040" (UID: "a4aaf26d-004b-4fe6-9896-dd0530fdd040"). InnerVolumeSpecName "kube-api-access-mz69s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:07:02 crc kubenswrapper[4702]: I1125 14:07:02.234520 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glancef364-account-delete-26dg2"] Nov 25 14:07:02 crc kubenswrapper[4702]: I1125 14:07:02.243261 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4aaf26d-004b-4fe6-9896-dd0530fdd040-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "a4aaf26d-004b-4fe6-9896-dd0530fdd040" (UID: "a4aaf26d-004b-4fe6-9896-dd0530fdd040"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 14:07:02 crc kubenswrapper[4702]: I1125 14:07:02.243285 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4aaf26d-004b-4fe6-9896-dd0530fdd040-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "a4aaf26d-004b-4fe6-9896-dd0530fdd040" (UID: "a4aaf26d-004b-4fe6-9896-dd0530fdd040"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:07:02 crc kubenswrapper[4702]: I1125 14:07:02.321039 4702 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a4aaf26d-004b-4fe6-9896-dd0530fdd040-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:02 crc kubenswrapper[4702]: I1125 14:07:02.321090 4702 reconciler_common.go:293] "Volume detached for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/a4aaf26d-004b-4fe6-9896-dd0530fdd040-openstack-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:02 crc kubenswrapper[4702]: I1125 14:07:02.321100 4702 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a4aaf26d-004b-4fe6-9896-dd0530fdd040-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:02 crc kubenswrapper[4702]: I1125 14:07:02.321111 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mz69s\" (UniqueName: \"kubernetes.io/projected/a4aaf26d-004b-4fe6-9896-dd0530fdd040-kube-api-access-mz69s\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:02 crc kubenswrapper[4702]: I1125 14:07:02.532656 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 14:07:02 crc kubenswrapper[4702]: I1125 14:07:02.542669 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 14:07:03 crc kubenswrapper[4702]: I1125 14:07:03.209777 4702 generic.go:334] "Generic (PLEG): container finished" podID="32b081d0-f397-4b90-bae2-8a398aa32084" containerID="e319cefabb880657e7e32ed663cba7b7067cd13a0d028da5e0f17f9faa7857f9" exitCode=0 Nov 25 14:07:03 crc kubenswrapper[4702]: I1125 14:07:03.209859 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancef364-account-delete-26dg2" event={"ID":"32b081d0-f397-4b90-bae2-8a398aa32084","Type":"ContainerDied","Data":"e319cefabb880657e7e32ed663cba7b7067cd13a0d028da5e0f17f9faa7857f9"} Nov 25 14:07:03 crc kubenswrapper[4702]: I1125 14:07:03.209888 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancef364-account-delete-26dg2" event={"ID":"32b081d0-f397-4b90-bae2-8a398aa32084","Type":"ContainerStarted","Data":"a2ec206f793f9e8f56cc9cbbc1c0d53fc0471d1133f9a5e65aaa560791fa5dfe"} Nov 25 14:07:03 crc kubenswrapper[4702]: I1125 14:07:03.811557 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4aaf26d-004b-4fe6-9896-dd0530fdd040" path="/var/lib/kubelet/pods/a4aaf26d-004b-4fe6-9896-dd0530fdd040/volumes" Nov 25 14:07:03 crc kubenswrapper[4702]: I1125 14:07:03.963450 4702 patch_prober.go:28] interesting pod/machine-config-daemon-qqr76 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 14:07:03 crc kubenswrapper[4702]: I1125 14:07:03.963575 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 14:07:04 crc kubenswrapper[4702]: I1125 14:07:04.559835 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancef364-account-delete-26dg2" Nov 25 14:07:04 crc kubenswrapper[4702]: I1125 14:07:04.652187 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-1" podUID="cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.0.100:9292/healthcheck\": read tcp 10.217.0.2:53422->10.217.0.100:9292: read: connection reset by peer" Nov 25 14:07:04 crc kubenswrapper[4702]: I1125 14:07:04.652219 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-1" podUID="cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.0.100:9292/healthcheck\": read tcp 10.217.0.2:53418->10.217.0.100:9292: read: connection reset by peer" Nov 25 14:07:04 crc kubenswrapper[4702]: I1125 14:07:04.681544 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32b081d0-f397-4b90-bae2-8a398aa32084-operator-scripts\") pod \"32b081d0-f397-4b90-bae2-8a398aa32084\" (UID: \"32b081d0-f397-4b90-bae2-8a398aa32084\") " Nov 25 14:07:04 crc kubenswrapper[4702]: I1125 14:07:04.681615 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bv42\" (UniqueName: \"kubernetes.io/projected/32b081d0-f397-4b90-bae2-8a398aa32084-kube-api-access-8bv42\") pod \"32b081d0-f397-4b90-bae2-8a398aa32084\" (UID: \"32b081d0-f397-4b90-bae2-8a398aa32084\") " Nov 25 14:07:04 crc kubenswrapper[4702]: I1125 14:07:04.682677 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32b081d0-f397-4b90-bae2-8a398aa32084-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "32b081d0-f397-4b90-bae2-8a398aa32084" (UID: "32b081d0-f397-4b90-bae2-8a398aa32084"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 14:07:04 crc kubenswrapper[4702]: I1125 14:07:04.688801 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32b081d0-f397-4b90-bae2-8a398aa32084-kube-api-access-8bv42" (OuterVolumeSpecName: "kube-api-access-8bv42") pod "32b081d0-f397-4b90-bae2-8a398aa32084" (UID: "32b081d0-f397-4b90-bae2-8a398aa32084"). InnerVolumeSpecName "kube-api-access-8bv42". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:07:04 crc kubenswrapper[4702]: I1125 14:07:04.783576 4702 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32b081d0-f397-4b90-bae2-8a398aa32084-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:04 crc kubenswrapper[4702]: I1125 14:07:04.783645 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bv42\" (UniqueName: \"kubernetes.io/projected/32b081d0-f397-4b90-bae2-8a398aa32084-kube-api-access-8bv42\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.022542 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.058822 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.199216 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-scripts\") pod \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.199311 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-config-data\") pod \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.199338 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3813201d-36d3-4599-9d57-d5b8a5b96f13-config-data\") pod \"3813201d-36d3-4599-9d57-d5b8a5b96f13\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.200030 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-var-locks-brick\") pod \"3813201d-36d3-4599-9d57-d5b8a5b96f13\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.200102 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-etc-iscsi\") pod \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.200138 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.200161 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-lib-modules\") pod \"3813201d-36d3-4599-9d57-d5b8a5b96f13\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.200217 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-sys\") pod \"3813201d-36d3-4599-9d57-d5b8a5b96f13\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.200303 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-dev\") pod \"3813201d-36d3-4599-9d57-d5b8a5b96f13\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.200344 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-run\") pod \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.200359 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.200380 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3813201d-36d3-4599-9d57-d5b8a5b96f13-logs\") pod \"3813201d-36d3-4599-9d57-d5b8a5b96f13\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.200412 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-lib-modules\") pod \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.200426 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-etc-nvme\") pod \"3813201d-36d3-4599-9d57-d5b8a5b96f13\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.200448 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"3813201d-36d3-4599-9d57-d5b8a5b96f13\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.200491 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-etc-nvme\") pod \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.200522 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-httpd-run\") pod \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.200547 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-etc-iscsi\") pod \"3813201d-36d3-4599-9d57-d5b8a5b96f13\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.200570 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-logs\") pod \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.200590 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-np7gd\" (UniqueName: \"kubernetes.io/projected/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-kube-api-access-np7gd\") pod \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.200608 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-run\") pod \"3813201d-36d3-4599-9d57-d5b8a5b96f13\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.200646 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-var-locks-brick\") pod \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.200678 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62nwp\" (UniqueName: \"kubernetes.io/projected/3813201d-36d3-4599-9d57-d5b8a5b96f13-kube-api-access-62nwp\") pod \"3813201d-36d3-4599-9d57-d5b8a5b96f13\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.200696 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3813201d-36d3-4599-9d57-d5b8a5b96f13-scripts\") pod \"3813201d-36d3-4599-9d57-d5b8a5b96f13\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.200719 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-sys\") pod \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.200741 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"3813201d-36d3-4599-9d57-d5b8a5b96f13\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.200764 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3813201d-36d3-4599-9d57-d5b8a5b96f13-httpd-run\") pod \"3813201d-36d3-4599-9d57-d5b8a5b96f13\" (UID: \"3813201d-36d3-4599-9d57-d5b8a5b96f13\") " Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.200792 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-dev\") pod \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\" (UID: \"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e\") " Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.201117 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-run" (OuterVolumeSpecName: "run") pod "3813201d-36d3-4599-9d57-d5b8a5b96f13" (UID: "3813201d-36d3-4599-9d57-d5b8a5b96f13"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.201137 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e" (UID: "cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.201208 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-sys" (OuterVolumeSpecName: "sys") pod "cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e" (UID: "cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.201250 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e" (UID: "cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.201191 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-dev" (OuterVolumeSpecName: "dev") pod "cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e" (UID: "cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.201278 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e" (UID: "cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.201294 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "3813201d-36d3-4599-9d57-d5b8a5b96f13" (UID: "3813201d-36d3-4599-9d57-d5b8a5b96f13"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.201433 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e" (UID: "cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.201469 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "3813201d-36d3-4599-9d57-d5b8a5b96f13" (UID: "3813201d-36d3-4599-9d57-d5b8a5b96f13"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.201453 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "3813201d-36d3-4599-9d57-d5b8a5b96f13" (UID: "3813201d-36d3-4599-9d57-d5b8a5b96f13"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.201528 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "3813201d-36d3-4599-9d57-d5b8a5b96f13" (UID: "3813201d-36d3-4599-9d57-d5b8a5b96f13"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.201588 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3813201d-36d3-4599-9d57-d5b8a5b96f13-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3813201d-36d3-4599-9d57-d5b8a5b96f13" (UID: "3813201d-36d3-4599-9d57-d5b8a5b96f13"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.201755 4702 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-sys\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.201777 4702 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3813201d-36d3-4599-9d57-d5b8a5b96f13-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.201791 4702 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-dev\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.201800 4702 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.201809 4702 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.201819 4702 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.201828 4702 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.201837 4702 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.201840 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3813201d-36d3-4599-9d57-d5b8a5b96f13-logs" (OuterVolumeSpecName: "logs") pod "3813201d-36d3-4599-9d57-d5b8a5b96f13" (UID: "3813201d-36d3-4599-9d57-d5b8a5b96f13"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.201846 4702 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.201901 4702 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.201914 4702 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-run\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.201926 4702 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.201959 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-sys" (OuterVolumeSpecName: "sys") pod "3813201d-36d3-4599-9d57-d5b8a5b96f13" (UID: "3813201d-36d3-4599-9d57-d5b8a5b96f13"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.201984 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-dev" (OuterVolumeSpecName: "dev") pod "3813201d-36d3-4599-9d57-d5b8a5b96f13" (UID: "3813201d-36d3-4599-9d57-d5b8a5b96f13"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.202010 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-run" (OuterVolumeSpecName: "run") pod "cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e" (UID: "cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.204161 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance-cache") pod "3813201d-36d3-4599-9d57-d5b8a5b96f13" (UID: "3813201d-36d3-4599-9d57-d5b8a5b96f13"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.205251 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3813201d-36d3-4599-9d57-d5b8a5b96f13-kube-api-access-62nwp" (OuterVolumeSpecName: "kube-api-access-62nwp") pod "3813201d-36d3-4599-9d57-d5b8a5b96f13" (UID: "3813201d-36d3-4599-9d57-d5b8a5b96f13"). InnerVolumeSpecName "kube-api-access-62nwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.205351 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3813201d-36d3-4599-9d57-d5b8a5b96f13-scripts" (OuterVolumeSpecName: "scripts") pod "3813201d-36d3-4599-9d57-d5b8a5b96f13" (UID: "3813201d-36d3-4599-9d57-d5b8a5b96f13"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.207276 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "3813201d-36d3-4599-9d57-d5b8a5b96f13" (UID: "3813201d-36d3-4599-9d57-d5b8a5b96f13"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.238533 4702 generic.go:334] "Generic (PLEG): container finished" podID="cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e" containerID="e8a739726777867d70cfc1d62be96bb994733c4ff18896df97aedbcbaa3b799a" exitCode=0 Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.238634 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e","Type":"ContainerDied","Data":"e8a739726777867d70cfc1d62be96bb994733c4ff18896df97aedbcbaa3b799a"} Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.238670 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e","Type":"ContainerDied","Data":"ec2c5568defe7892250e025ca66b0f2f9cedc9130385a4539250095907535e42"} Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.238690 4702 scope.go:117] "RemoveContainer" containerID="e8a739726777867d70cfc1d62be96bb994733c4ff18896df97aedbcbaa3b799a" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.238708 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.243197 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancef364-account-delete-26dg2" event={"ID":"32b081d0-f397-4b90-bae2-8a398aa32084","Type":"ContainerDied","Data":"a2ec206f793f9e8f56cc9cbbc1c0d53fc0471d1133f9a5e65aaa560791fa5dfe"} Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.243292 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2ec206f793f9e8f56cc9cbbc1c0d53fc0471d1133f9a5e65aaa560791fa5dfe" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.243292 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancef364-account-delete-26dg2" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.245287 4702 generic.go:334] "Generic (PLEG): container finished" podID="3813201d-36d3-4599-9d57-d5b8a5b96f13" containerID="ea926cd6e89698d0a6c7550aeffd0d571d9babf3ba5b35e0320cfe15a608addd" exitCode=0 Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.245333 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"3813201d-36d3-4599-9d57-d5b8a5b96f13","Type":"ContainerDied","Data":"ea926cd6e89698d0a6c7550aeffd0d571d9babf3ba5b35e0320cfe15a608addd"} Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.245365 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"3813201d-36d3-4599-9d57-d5b8a5b96f13","Type":"ContainerDied","Data":"7b8d7f358de9dae84426267a1aa30c8a9b290374a205102d58df06a44735fcc2"} Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.245428 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.303967 4702 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-dev\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.304005 4702 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-run\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.304014 4702 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3813201d-36d3-4599-9d57-d5b8a5b96f13-logs\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.304047 4702 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.304058 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62nwp\" (UniqueName: \"kubernetes.io/projected/3813201d-36d3-4599-9d57-d5b8a5b96f13-kube-api-access-62nwp\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.304070 4702 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3813201d-36d3-4599-9d57-d5b8a5b96f13-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.304084 4702 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.304095 4702 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3813201d-36d3-4599-9d57-d5b8a5b96f13-sys\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.316741 4702 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.317272 4702 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.361907 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-scripts" (OuterVolumeSpecName: "scripts") pod "cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e" (UID: "cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.378508 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e" (UID: "cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.383378 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-config-data" (OuterVolumeSpecName: "config-data") pod "cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e" (UID: "cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.405576 4702 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.405608 4702 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.405618 4702 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.405627 4702 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.405636 4702 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.410635 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-kube-api-access-np7gd" (OuterVolumeSpecName: "kube-api-access-np7gd") pod "cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e" (UID: "cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e"). InnerVolumeSpecName "kube-api-access-np7gd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.411749 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance-cache") pod "cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e" (UID: "cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.412451 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e" (UID: "cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.445560 4702 scope.go:117] "RemoveContainer" containerID="41f52290460e870d5a774f97567d5922d97d5a39d98b2a7853314fb70f9bf11d" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.457690 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-logs" (OuterVolumeSpecName: "logs") pod "cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e" (UID: "cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.465345 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3813201d-36d3-4599-9d57-d5b8a5b96f13-config-data" (OuterVolumeSpecName: "config-data") pod "3813201d-36d3-4599-9d57-d5b8a5b96f13" (UID: "3813201d-36d3-4599-9d57-d5b8a5b96f13"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.467330 4702 scope.go:117] "RemoveContainer" containerID="e8a739726777867d70cfc1d62be96bb994733c4ff18896df97aedbcbaa3b799a" Nov 25 14:07:05 crc kubenswrapper[4702]: E1125 14:07:05.470089 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8a739726777867d70cfc1d62be96bb994733c4ff18896df97aedbcbaa3b799a\": container with ID starting with e8a739726777867d70cfc1d62be96bb994733c4ff18896df97aedbcbaa3b799a not found: ID does not exist" containerID="e8a739726777867d70cfc1d62be96bb994733c4ff18896df97aedbcbaa3b799a" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.470186 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8a739726777867d70cfc1d62be96bb994733c4ff18896df97aedbcbaa3b799a"} err="failed to get container status \"e8a739726777867d70cfc1d62be96bb994733c4ff18896df97aedbcbaa3b799a\": rpc error: code = NotFound desc = could not find container \"e8a739726777867d70cfc1d62be96bb994733c4ff18896df97aedbcbaa3b799a\": container with ID starting with e8a739726777867d70cfc1d62be96bb994733c4ff18896df97aedbcbaa3b799a not found: ID does not exist" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.470287 4702 scope.go:117] "RemoveContainer" containerID="41f52290460e870d5a774f97567d5922d97d5a39d98b2a7853314fb70f9bf11d" Nov 25 14:07:05 crc kubenswrapper[4702]: E1125 14:07:05.470842 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41f52290460e870d5a774f97567d5922d97d5a39d98b2a7853314fb70f9bf11d\": container with ID starting with 41f52290460e870d5a774f97567d5922d97d5a39d98b2a7853314fb70f9bf11d not found: ID does not exist" containerID="41f52290460e870d5a774f97567d5922d97d5a39d98b2a7853314fb70f9bf11d" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.470891 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41f52290460e870d5a774f97567d5922d97d5a39d98b2a7853314fb70f9bf11d"} err="failed to get container status \"41f52290460e870d5a774f97567d5922d97d5a39d98b2a7853314fb70f9bf11d\": rpc error: code = NotFound desc = could not find container \"41f52290460e870d5a774f97567d5922d97d5a39d98b2a7853314fb70f9bf11d\": container with ID starting with 41f52290460e870d5a774f97567d5922d97d5a39d98b2a7853314fb70f9bf11d not found: ID does not exist" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.470922 4702 scope.go:117] "RemoveContainer" containerID="ea926cd6e89698d0a6c7550aeffd0d571d9babf3ba5b35e0320cfe15a608addd" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.494857 4702 scope.go:117] "RemoveContainer" containerID="80e2dd0a5e92cadc9c3dfad4aa2e542f1c5dfa142baea0466c9c4ccb1c41b76a" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.507120 4702 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.507176 4702 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.507195 4702 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-logs\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.507210 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-np7gd\" (UniqueName: \"kubernetes.io/projected/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e-kube-api-access-np7gd\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.507227 4702 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3813201d-36d3-4599-9d57-d5b8a5b96f13-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.521029 4702 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.522186 4702 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.553429 4702 scope.go:117] "RemoveContainer" containerID="ea926cd6e89698d0a6c7550aeffd0d571d9babf3ba5b35e0320cfe15a608addd" Nov 25 14:07:05 crc kubenswrapper[4702]: E1125 14:07:05.554050 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea926cd6e89698d0a6c7550aeffd0d571d9babf3ba5b35e0320cfe15a608addd\": container with ID starting with ea926cd6e89698d0a6c7550aeffd0d571d9babf3ba5b35e0320cfe15a608addd not found: ID does not exist" containerID="ea926cd6e89698d0a6c7550aeffd0d571d9babf3ba5b35e0320cfe15a608addd" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.554153 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea926cd6e89698d0a6c7550aeffd0d571d9babf3ba5b35e0320cfe15a608addd"} err="failed to get container status \"ea926cd6e89698d0a6c7550aeffd0d571d9babf3ba5b35e0320cfe15a608addd\": rpc error: code = NotFound desc = could not find container \"ea926cd6e89698d0a6c7550aeffd0d571d9babf3ba5b35e0320cfe15a608addd\": container with ID starting with ea926cd6e89698d0a6c7550aeffd0d571d9babf3ba5b35e0320cfe15a608addd not found: ID does not exist" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.554194 4702 scope.go:117] "RemoveContainer" containerID="80e2dd0a5e92cadc9c3dfad4aa2e542f1c5dfa142baea0466c9c4ccb1c41b76a" Nov 25 14:07:05 crc kubenswrapper[4702]: E1125 14:07:05.555290 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80e2dd0a5e92cadc9c3dfad4aa2e542f1c5dfa142baea0466c9c4ccb1c41b76a\": container with ID starting with 80e2dd0a5e92cadc9c3dfad4aa2e542f1c5dfa142baea0466c9c4ccb1c41b76a not found: ID does not exist" containerID="80e2dd0a5e92cadc9c3dfad4aa2e542f1c5dfa142baea0466c9c4ccb1c41b76a" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.555339 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80e2dd0a5e92cadc9c3dfad4aa2e542f1c5dfa142baea0466c9c4ccb1c41b76a"} err="failed to get container status \"80e2dd0a5e92cadc9c3dfad4aa2e542f1c5dfa142baea0466c9c4ccb1c41b76a\": rpc error: code = NotFound desc = could not find container \"80e2dd0a5e92cadc9c3dfad4aa2e542f1c5dfa142baea0466c9c4ccb1c41b76a\": container with ID starting with 80e2dd0a5e92cadc9c3dfad4aa2e542f1c5dfa142baea0466c9c4ccb1c41b76a not found: ID does not exist" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.588146 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.601751 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.607908 4702 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.607950 4702 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.620414 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.625647 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.812878 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3813201d-36d3-4599-9d57-d5b8a5b96f13" path="/var/lib/kubelet/pods/3813201d-36d3-4599-9d57-d5b8a5b96f13/volumes" Nov 25 14:07:05 crc kubenswrapper[4702]: I1125 14:07:05.813726 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e" path="/var/lib/kubelet/pods/cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e/volumes" Nov 25 14:07:06 crc kubenswrapper[4702]: I1125 14:07:06.443536 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-9t6bx"] Nov 25 14:07:06 crc kubenswrapper[4702]: I1125 14:07:06.452477 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-9t6bx"] Nov 25 14:07:06 crc kubenswrapper[4702]: I1125 14:07:06.458518 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glancef364-account-delete-26dg2"] Nov 25 14:07:06 crc kubenswrapper[4702]: I1125 14:07:06.464325 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glancef364-account-delete-26dg2"] Nov 25 14:07:06 crc kubenswrapper[4702]: I1125 14:07:06.470423 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-f364-account-create-update-xpwg6"] Nov 25 14:07:06 crc kubenswrapper[4702]: I1125 14:07:06.476739 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-f364-account-create-update-xpwg6"] Nov 25 14:07:07 crc kubenswrapper[4702]: I1125 14:07:07.811125 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28c2c617-0d70-4223-a7b3-0043eb3b7d87" path="/var/lib/kubelet/pods/28c2c617-0d70-4223-a7b3-0043eb3b7d87/volumes" Nov 25 14:07:07 crc kubenswrapper[4702]: I1125 14:07:07.811882 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30d08778-c048-4b5f-b0d5-d018e4af5b1f" path="/var/lib/kubelet/pods/30d08778-c048-4b5f-b0d5-d018e4af5b1f/volumes" Nov 25 14:07:07 crc kubenswrapper[4702]: I1125 14:07:07.812380 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32b081d0-f397-4b90-bae2-8a398aa32084" path="/var/lib/kubelet/pods/32b081d0-f397-4b90-bae2-8a398aa32084/volumes" Nov 25 14:07:07 crc kubenswrapper[4702]: I1125 14:07:07.955563 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-pth7k"] Nov 25 14:07:07 crc kubenswrapper[4702]: E1125 14:07:07.955850 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3813201d-36d3-4599-9d57-d5b8a5b96f13" containerName="glance-httpd" Nov 25 14:07:07 crc kubenswrapper[4702]: I1125 14:07:07.955863 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="3813201d-36d3-4599-9d57-d5b8a5b96f13" containerName="glance-httpd" Nov 25 14:07:07 crc kubenswrapper[4702]: E1125 14:07:07.955877 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e" containerName="glance-httpd" Nov 25 14:07:07 crc kubenswrapper[4702]: I1125 14:07:07.955883 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e" containerName="glance-httpd" Nov 25 14:07:07 crc kubenswrapper[4702]: E1125 14:07:07.955901 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3813201d-36d3-4599-9d57-d5b8a5b96f13" containerName="glance-log" Nov 25 14:07:07 crc kubenswrapper[4702]: I1125 14:07:07.955908 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="3813201d-36d3-4599-9d57-d5b8a5b96f13" containerName="glance-log" Nov 25 14:07:07 crc kubenswrapper[4702]: E1125 14:07:07.955923 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4aaf26d-004b-4fe6-9896-dd0530fdd040" containerName="openstackclient" Nov 25 14:07:07 crc kubenswrapper[4702]: I1125 14:07:07.955929 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4aaf26d-004b-4fe6-9896-dd0530fdd040" containerName="openstackclient" Nov 25 14:07:07 crc kubenswrapper[4702]: E1125 14:07:07.955942 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e" containerName="glance-log" Nov 25 14:07:07 crc kubenswrapper[4702]: I1125 14:07:07.955948 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e" containerName="glance-log" Nov 25 14:07:07 crc kubenswrapper[4702]: E1125 14:07:07.955958 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32b081d0-f397-4b90-bae2-8a398aa32084" containerName="mariadb-account-delete" Nov 25 14:07:07 crc kubenswrapper[4702]: I1125 14:07:07.955965 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="32b081d0-f397-4b90-bae2-8a398aa32084" containerName="mariadb-account-delete" Nov 25 14:07:07 crc kubenswrapper[4702]: I1125 14:07:07.956298 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="32b081d0-f397-4b90-bae2-8a398aa32084" containerName="mariadb-account-delete" Nov 25 14:07:07 crc kubenswrapper[4702]: I1125 14:07:07.956309 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4aaf26d-004b-4fe6-9896-dd0530fdd040" containerName="openstackclient" Nov 25 14:07:07 crc kubenswrapper[4702]: I1125 14:07:07.956319 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="3813201d-36d3-4599-9d57-d5b8a5b96f13" containerName="glance-httpd" Nov 25 14:07:07 crc kubenswrapper[4702]: I1125 14:07:07.956328 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e" containerName="glance-log" Nov 25 14:07:07 crc kubenswrapper[4702]: I1125 14:07:07.956336 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="3813201d-36d3-4599-9d57-d5b8a5b96f13" containerName="glance-log" Nov 25 14:07:07 crc kubenswrapper[4702]: I1125 14:07:07.956342 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbb1bc07-9697-4f91-bfa2-b00fcb85fe5e" containerName="glance-httpd" Nov 25 14:07:07 crc kubenswrapper[4702]: I1125 14:07:07.956895 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-pth7k" Nov 25 14:07:07 crc kubenswrapper[4702]: I1125 14:07:07.972545 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-pth7k"] Nov 25 14:07:07 crc kubenswrapper[4702]: I1125 14:07:07.980494 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-faf0-account-create-update-p4425"] Nov 25 14:07:07 crc kubenswrapper[4702]: I1125 14:07:07.982178 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-faf0-account-create-update-p4425" Nov 25 14:07:07 crc kubenswrapper[4702]: I1125 14:07:07.984999 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 25 14:07:08 crc kubenswrapper[4702]: I1125 14:07:08.011489 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-faf0-account-create-update-p4425"] Nov 25 14:07:08 crc kubenswrapper[4702]: I1125 14:07:08.052351 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65a7b32a-6d99-4e03-92e7-a2d4f9e44441-operator-scripts\") pod \"glance-faf0-account-create-update-p4425\" (UID: \"65a7b32a-6d99-4e03-92e7-a2d4f9e44441\") " pod="glance-kuttl-tests/glance-faf0-account-create-update-p4425" Nov 25 14:07:08 crc kubenswrapper[4702]: I1125 14:07:08.052629 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73b0c1bd-c025-419a-8a97-03523b6844b2-operator-scripts\") pod \"glance-db-create-pth7k\" (UID: \"73b0c1bd-c025-419a-8a97-03523b6844b2\") " pod="glance-kuttl-tests/glance-db-create-pth7k" Nov 25 14:07:08 crc kubenswrapper[4702]: I1125 14:07:08.052728 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7flw\" (UniqueName: \"kubernetes.io/projected/73b0c1bd-c025-419a-8a97-03523b6844b2-kube-api-access-t7flw\") pod \"glance-db-create-pth7k\" (UID: \"73b0c1bd-c025-419a-8a97-03523b6844b2\") " pod="glance-kuttl-tests/glance-db-create-pth7k" Nov 25 14:07:08 crc kubenswrapper[4702]: I1125 14:07:08.052984 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dd9c\" (UniqueName: \"kubernetes.io/projected/65a7b32a-6d99-4e03-92e7-a2d4f9e44441-kube-api-access-4dd9c\") pod \"glance-faf0-account-create-update-p4425\" (UID: \"65a7b32a-6d99-4e03-92e7-a2d4f9e44441\") " pod="glance-kuttl-tests/glance-faf0-account-create-update-p4425" Nov 25 14:07:08 crc kubenswrapper[4702]: I1125 14:07:08.154330 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73b0c1bd-c025-419a-8a97-03523b6844b2-operator-scripts\") pod \"glance-db-create-pth7k\" (UID: \"73b0c1bd-c025-419a-8a97-03523b6844b2\") " pod="glance-kuttl-tests/glance-db-create-pth7k" Nov 25 14:07:08 crc kubenswrapper[4702]: I1125 14:07:08.154420 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7flw\" (UniqueName: \"kubernetes.io/projected/73b0c1bd-c025-419a-8a97-03523b6844b2-kube-api-access-t7flw\") pod \"glance-db-create-pth7k\" (UID: \"73b0c1bd-c025-419a-8a97-03523b6844b2\") " pod="glance-kuttl-tests/glance-db-create-pth7k" Nov 25 14:07:08 crc kubenswrapper[4702]: I1125 14:07:08.154499 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dd9c\" (UniqueName: \"kubernetes.io/projected/65a7b32a-6d99-4e03-92e7-a2d4f9e44441-kube-api-access-4dd9c\") pod \"glance-faf0-account-create-update-p4425\" (UID: \"65a7b32a-6d99-4e03-92e7-a2d4f9e44441\") " pod="glance-kuttl-tests/glance-faf0-account-create-update-p4425" Nov 25 14:07:08 crc kubenswrapper[4702]: I1125 14:07:08.154577 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65a7b32a-6d99-4e03-92e7-a2d4f9e44441-operator-scripts\") pod \"glance-faf0-account-create-update-p4425\" (UID: \"65a7b32a-6d99-4e03-92e7-a2d4f9e44441\") " pod="glance-kuttl-tests/glance-faf0-account-create-update-p4425" Nov 25 14:07:08 crc kubenswrapper[4702]: I1125 14:07:08.155478 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73b0c1bd-c025-419a-8a97-03523b6844b2-operator-scripts\") pod \"glance-db-create-pth7k\" (UID: \"73b0c1bd-c025-419a-8a97-03523b6844b2\") " pod="glance-kuttl-tests/glance-db-create-pth7k" Nov 25 14:07:08 crc kubenswrapper[4702]: I1125 14:07:08.155928 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65a7b32a-6d99-4e03-92e7-a2d4f9e44441-operator-scripts\") pod \"glance-faf0-account-create-update-p4425\" (UID: \"65a7b32a-6d99-4e03-92e7-a2d4f9e44441\") " pod="glance-kuttl-tests/glance-faf0-account-create-update-p4425" Nov 25 14:07:08 crc kubenswrapper[4702]: I1125 14:07:08.176077 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dd9c\" (UniqueName: \"kubernetes.io/projected/65a7b32a-6d99-4e03-92e7-a2d4f9e44441-kube-api-access-4dd9c\") pod \"glance-faf0-account-create-update-p4425\" (UID: \"65a7b32a-6d99-4e03-92e7-a2d4f9e44441\") " pod="glance-kuttl-tests/glance-faf0-account-create-update-p4425" Nov 25 14:07:08 crc kubenswrapper[4702]: I1125 14:07:08.176913 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7flw\" (UniqueName: \"kubernetes.io/projected/73b0c1bd-c025-419a-8a97-03523b6844b2-kube-api-access-t7flw\") pod \"glance-db-create-pth7k\" (UID: \"73b0c1bd-c025-419a-8a97-03523b6844b2\") " pod="glance-kuttl-tests/glance-db-create-pth7k" Nov 25 14:07:08 crc kubenswrapper[4702]: I1125 14:07:08.277158 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-pth7k" Nov 25 14:07:08 crc kubenswrapper[4702]: I1125 14:07:08.306751 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-faf0-account-create-update-p4425" Nov 25 14:07:08 crc kubenswrapper[4702]: I1125 14:07:08.719093 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-pth7k"] Nov 25 14:07:08 crc kubenswrapper[4702]: I1125 14:07:08.779585 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-faf0-account-create-update-p4425"] Nov 25 14:07:08 crc kubenswrapper[4702]: W1125 14:07:08.791032 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65a7b32a_6d99_4e03_92e7_a2d4f9e44441.slice/crio-4c098df65d42507488422b82e3d1615f7cb29b24494d3d3cc5437fc7d6b81eec WatchSource:0}: Error finding container 4c098df65d42507488422b82e3d1615f7cb29b24494d3d3cc5437fc7d6b81eec: Status 404 returned error can't find the container with id 4c098df65d42507488422b82e3d1615f7cb29b24494d3d3cc5437fc7d6b81eec Nov 25 14:07:09 crc kubenswrapper[4702]: E1125 14:07:09.269639 4702 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65a7b32a_6d99_4e03_92e7_a2d4f9e44441.slice/crio-f336a0efc3b1011c580f7efb9b0362e23c98ae426a5226b3ecb6e71a1275a841.scope\": RecentStats: unable to find data in memory cache]" Nov 25 14:07:09 crc kubenswrapper[4702]: I1125 14:07:09.285557 4702 generic.go:334] "Generic (PLEG): container finished" podID="65a7b32a-6d99-4e03-92e7-a2d4f9e44441" containerID="f336a0efc3b1011c580f7efb9b0362e23c98ae426a5226b3ecb6e71a1275a841" exitCode=0 Nov 25 14:07:09 crc kubenswrapper[4702]: I1125 14:07:09.285627 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-faf0-account-create-update-p4425" event={"ID":"65a7b32a-6d99-4e03-92e7-a2d4f9e44441","Type":"ContainerDied","Data":"f336a0efc3b1011c580f7efb9b0362e23c98ae426a5226b3ecb6e71a1275a841"} Nov 25 14:07:09 crc kubenswrapper[4702]: I1125 14:07:09.285662 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-faf0-account-create-update-p4425" event={"ID":"65a7b32a-6d99-4e03-92e7-a2d4f9e44441","Type":"ContainerStarted","Data":"4c098df65d42507488422b82e3d1615f7cb29b24494d3d3cc5437fc7d6b81eec"} Nov 25 14:07:09 crc kubenswrapper[4702]: I1125 14:07:09.321868 4702 generic.go:334] "Generic (PLEG): container finished" podID="73b0c1bd-c025-419a-8a97-03523b6844b2" containerID="a7bc21f37028e958666ea2c6c8d71a826c4871f9071098876c4467c088b82dc0" exitCode=0 Nov 25 14:07:09 crc kubenswrapper[4702]: I1125 14:07:09.321945 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-pth7k" event={"ID":"73b0c1bd-c025-419a-8a97-03523b6844b2","Type":"ContainerDied","Data":"a7bc21f37028e958666ea2c6c8d71a826c4871f9071098876c4467c088b82dc0"} Nov 25 14:07:09 crc kubenswrapper[4702]: I1125 14:07:09.321986 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-pth7k" event={"ID":"73b0c1bd-c025-419a-8a97-03523b6844b2","Type":"ContainerStarted","Data":"62590c2658aa4047dd401a5e16f11bb65a7b5a34c76ab1cedcf1c29c0dbb17fb"} Nov 25 14:07:10 crc kubenswrapper[4702]: I1125 14:07:10.705441 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-pth7k" Nov 25 14:07:10 crc kubenswrapper[4702]: I1125 14:07:10.788135 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-faf0-account-create-update-p4425" Nov 25 14:07:10 crc kubenswrapper[4702]: I1125 14:07:10.903850 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7flw\" (UniqueName: \"kubernetes.io/projected/73b0c1bd-c025-419a-8a97-03523b6844b2-kube-api-access-t7flw\") pod \"73b0c1bd-c025-419a-8a97-03523b6844b2\" (UID: \"73b0c1bd-c025-419a-8a97-03523b6844b2\") " Nov 25 14:07:10 crc kubenswrapper[4702]: I1125 14:07:10.904138 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65a7b32a-6d99-4e03-92e7-a2d4f9e44441-operator-scripts\") pod \"65a7b32a-6d99-4e03-92e7-a2d4f9e44441\" (UID: \"65a7b32a-6d99-4e03-92e7-a2d4f9e44441\") " Nov 25 14:07:10 crc kubenswrapper[4702]: I1125 14:07:10.904215 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dd9c\" (UniqueName: \"kubernetes.io/projected/65a7b32a-6d99-4e03-92e7-a2d4f9e44441-kube-api-access-4dd9c\") pod \"65a7b32a-6d99-4e03-92e7-a2d4f9e44441\" (UID: \"65a7b32a-6d99-4e03-92e7-a2d4f9e44441\") " Nov 25 14:07:10 crc kubenswrapper[4702]: I1125 14:07:10.904267 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73b0c1bd-c025-419a-8a97-03523b6844b2-operator-scripts\") pod \"73b0c1bd-c025-419a-8a97-03523b6844b2\" (UID: \"73b0c1bd-c025-419a-8a97-03523b6844b2\") " Nov 25 14:07:10 crc kubenswrapper[4702]: I1125 14:07:10.905276 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73b0c1bd-c025-419a-8a97-03523b6844b2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "73b0c1bd-c025-419a-8a97-03523b6844b2" (UID: "73b0c1bd-c025-419a-8a97-03523b6844b2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 14:07:10 crc kubenswrapper[4702]: I1125 14:07:10.905671 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65a7b32a-6d99-4e03-92e7-a2d4f9e44441-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "65a7b32a-6d99-4e03-92e7-a2d4f9e44441" (UID: "65a7b32a-6d99-4e03-92e7-a2d4f9e44441"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 14:07:10 crc kubenswrapper[4702]: I1125 14:07:10.917143 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65a7b32a-6d99-4e03-92e7-a2d4f9e44441-kube-api-access-4dd9c" (OuterVolumeSpecName: "kube-api-access-4dd9c") pod "65a7b32a-6d99-4e03-92e7-a2d4f9e44441" (UID: "65a7b32a-6d99-4e03-92e7-a2d4f9e44441"). InnerVolumeSpecName "kube-api-access-4dd9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:07:10 crc kubenswrapper[4702]: I1125 14:07:10.917215 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73b0c1bd-c025-419a-8a97-03523b6844b2-kube-api-access-t7flw" (OuterVolumeSpecName: "kube-api-access-t7flw") pod "73b0c1bd-c025-419a-8a97-03523b6844b2" (UID: "73b0c1bd-c025-419a-8a97-03523b6844b2"). InnerVolumeSpecName "kube-api-access-t7flw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:07:11 crc kubenswrapper[4702]: I1125 14:07:11.007010 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dd9c\" (UniqueName: \"kubernetes.io/projected/65a7b32a-6d99-4e03-92e7-a2d4f9e44441-kube-api-access-4dd9c\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:11 crc kubenswrapper[4702]: I1125 14:07:11.007049 4702 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73b0c1bd-c025-419a-8a97-03523b6844b2-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:11 crc kubenswrapper[4702]: I1125 14:07:11.007063 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7flw\" (UniqueName: \"kubernetes.io/projected/73b0c1bd-c025-419a-8a97-03523b6844b2-kube-api-access-t7flw\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:11 crc kubenswrapper[4702]: I1125 14:07:11.007079 4702 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65a7b32a-6d99-4e03-92e7-a2d4f9e44441-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:11 crc kubenswrapper[4702]: I1125 14:07:11.346157 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-faf0-account-create-update-p4425" event={"ID":"65a7b32a-6d99-4e03-92e7-a2d4f9e44441","Type":"ContainerDied","Data":"4c098df65d42507488422b82e3d1615f7cb29b24494d3d3cc5437fc7d6b81eec"} Nov 25 14:07:11 crc kubenswrapper[4702]: I1125 14:07:11.346218 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c098df65d42507488422b82e3d1615f7cb29b24494d3d3cc5437fc7d6b81eec" Nov 25 14:07:11 crc kubenswrapper[4702]: I1125 14:07:11.346294 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-faf0-account-create-update-p4425" Nov 25 14:07:11 crc kubenswrapper[4702]: I1125 14:07:11.348065 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-pth7k" event={"ID":"73b0c1bd-c025-419a-8a97-03523b6844b2","Type":"ContainerDied","Data":"62590c2658aa4047dd401a5e16f11bb65a7b5a34c76ab1cedcf1c29c0dbb17fb"} Nov 25 14:07:11 crc kubenswrapper[4702]: I1125 14:07:11.348123 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="62590c2658aa4047dd401a5e16f11bb65a7b5a34c76ab1cedcf1c29c0dbb17fb" Nov 25 14:07:11 crc kubenswrapper[4702]: I1125 14:07:11.348185 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-pth7k" Nov 25 14:07:13 crc kubenswrapper[4702]: I1125 14:07:13.206672 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-ldqjq"] Nov 25 14:07:13 crc kubenswrapper[4702]: E1125 14:07:13.207572 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73b0c1bd-c025-419a-8a97-03523b6844b2" containerName="mariadb-database-create" Nov 25 14:07:13 crc kubenswrapper[4702]: I1125 14:07:13.207590 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="73b0c1bd-c025-419a-8a97-03523b6844b2" containerName="mariadb-database-create" Nov 25 14:07:13 crc kubenswrapper[4702]: E1125 14:07:13.207606 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65a7b32a-6d99-4e03-92e7-a2d4f9e44441" containerName="mariadb-account-create-update" Nov 25 14:07:13 crc kubenswrapper[4702]: I1125 14:07:13.207614 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="65a7b32a-6d99-4e03-92e7-a2d4f9e44441" containerName="mariadb-account-create-update" Nov 25 14:07:13 crc kubenswrapper[4702]: I1125 14:07:13.207759 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="65a7b32a-6d99-4e03-92e7-a2d4f9e44441" containerName="mariadb-account-create-update" Nov 25 14:07:13 crc kubenswrapper[4702]: I1125 14:07:13.207781 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="73b0c1bd-c025-419a-8a97-03523b6844b2" containerName="mariadb-database-create" Nov 25 14:07:13 crc kubenswrapper[4702]: I1125 14:07:13.208627 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-ldqjq" Nov 25 14:07:13 crc kubenswrapper[4702]: I1125 14:07:13.211419 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Nov 25 14:07:13 crc kubenswrapper[4702]: I1125 14:07:13.211694 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-kjx4l" Nov 25 14:07:13 crc kubenswrapper[4702]: I1125 14:07:13.224744 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-ldqjq"] Nov 25 14:07:13 crc kubenswrapper[4702]: I1125 14:07:13.276969 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6cf87935-5546-46f6-b8bb-f254e611faa7-db-sync-config-data\") pod \"glance-db-sync-ldqjq\" (UID: \"6cf87935-5546-46f6-b8bb-f254e611faa7\") " pod="glance-kuttl-tests/glance-db-sync-ldqjq" Nov 25 14:07:13 crc kubenswrapper[4702]: I1125 14:07:13.277051 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cf87935-5546-46f6-b8bb-f254e611faa7-combined-ca-bundle\") pod \"glance-db-sync-ldqjq\" (UID: \"6cf87935-5546-46f6-b8bb-f254e611faa7\") " pod="glance-kuttl-tests/glance-db-sync-ldqjq" Nov 25 14:07:13 crc kubenswrapper[4702]: I1125 14:07:13.277132 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cf87935-5546-46f6-b8bb-f254e611faa7-config-data\") pod \"glance-db-sync-ldqjq\" (UID: \"6cf87935-5546-46f6-b8bb-f254e611faa7\") " pod="glance-kuttl-tests/glance-db-sync-ldqjq" Nov 25 14:07:13 crc kubenswrapper[4702]: I1125 14:07:13.277171 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 25 14:07:13 crc kubenswrapper[4702]: I1125 14:07:13.277489 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwwnq\" (UniqueName: \"kubernetes.io/projected/6cf87935-5546-46f6-b8bb-f254e611faa7-kube-api-access-nwwnq\") pod \"glance-db-sync-ldqjq\" (UID: \"6cf87935-5546-46f6-b8bb-f254e611faa7\") " pod="glance-kuttl-tests/glance-db-sync-ldqjq" Nov 25 14:07:13 crc kubenswrapper[4702]: I1125 14:07:13.378885 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6cf87935-5546-46f6-b8bb-f254e611faa7-db-sync-config-data\") pod \"glance-db-sync-ldqjq\" (UID: \"6cf87935-5546-46f6-b8bb-f254e611faa7\") " pod="glance-kuttl-tests/glance-db-sync-ldqjq" Nov 25 14:07:13 crc kubenswrapper[4702]: I1125 14:07:13.378965 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cf87935-5546-46f6-b8bb-f254e611faa7-combined-ca-bundle\") pod \"glance-db-sync-ldqjq\" (UID: \"6cf87935-5546-46f6-b8bb-f254e611faa7\") " pod="glance-kuttl-tests/glance-db-sync-ldqjq" Nov 25 14:07:13 crc kubenswrapper[4702]: I1125 14:07:13.379048 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cf87935-5546-46f6-b8bb-f254e611faa7-config-data\") pod \"glance-db-sync-ldqjq\" (UID: \"6cf87935-5546-46f6-b8bb-f254e611faa7\") " pod="glance-kuttl-tests/glance-db-sync-ldqjq" Nov 25 14:07:13 crc kubenswrapper[4702]: I1125 14:07:13.379110 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwwnq\" (UniqueName: \"kubernetes.io/projected/6cf87935-5546-46f6-b8bb-f254e611faa7-kube-api-access-nwwnq\") pod \"glance-db-sync-ldqjq\" (UID: \"6cf87935-5546-46f6-b8bb-f254e611faa7\") " pod="glance-kuttl-tests/glance-db-sync-ldqjq" Nov 25 14:07:13 crc kubenswrapper[4702]: I1125 14:07:13.385503 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6cf87935-5546-46f6-b8bb-f254e611faa7-db-sync-config-data\") pod \"glance-db-sync-ldqjq\" (UID: \"6cf87935-5546-46f6-b8bb-f254e611faa7\") " pod="glance-kuttl-tests/glance-db-sync-ldqjq" Nov 25 14:07:13 crc kubenswrapper[4702]: I1125 14:07:13.385548 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cf87935-5546-46f6-b8bb-f254e611faa7-combined-ca-bundle\") pod \"glance-db-sync-ldqjq\" (UID: \"6cf87935-5546-46f6-b8bb-f254e611faa7\") " pod="glance-kuttl-tests/glance-db-sync-ldqjq" Nov 25 14:07:13 crc kubenswrapper[4702]: I1125 14:07:13.394634 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cf87935-5546-46f6-b8bb-f254e611faa7-config-data\") pod \"glance-db-sync-ldqjq\" (UID: \"6cf87935-5546-46f6-b8bb-f254e611faa7\") " pod="glance-kuttl-tests/glance-db-sync-ldqjq" Nov 25 14:07:13 crc kubenswrapper[4702]: I1125 14:07:13.395286 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwwnq\" (UniqueName: \"kubernetes.io/projected/6cf87935-5546-46f6-b8bb-f254e611faa7-kube-api-access-nwwnq\") pod \"glance-db-sync-ldqjq\" (UID: \"6cf87935-5546-46f6-b8bb-f254e611faa7\") " pod="glance-kuttl-tests/glance-db-sync-ldqjq" Nov 25 14:07:13 crc kubenswrapper[4702]: I1125 14:07:13.586470 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-ldqjq" Nov 25 14:07:14 crc kubenswrapper[4702]: I1125 14:07:14.071114 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-ldqjq"] Nov 25 14:07:14 crc kubenswrapper[4702]: I1125 14:07:14.381609 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-ldqjq" event={"ID":"6cf87935-5546-46f6-b8bb-f254e611faa7","Type":"ContainerStarted","Data":"69c8b54d60f1392c6f766827bd599cd038ec067a45741ad2f09fb2c727b687fe"} Nov 25 14:07:15 crc kubenswrapper[4702]: I1125 14:07:15.392340 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-ldqjq" event={"ID":"6cf87935-5546-46f6-b8bb-f254e611faa7","Type":"ContainerStarted","Data":"2af2837b46377b1533459d4a0613c8e64b3a32f54338f03ae72128307b90e344"} Nov 25 14:07:15 crc kubenswrapper[4702]: I1125 14:07:15.413202 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-ldqjq" podStartSLOduration=2.413176955 podStartE2EDuration="2.413176955s" podCreationTimestamp="2025-11-25 14:07:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 14:07:15.409488824 +0000 UTC m=+1244.572139274" watchObservedRunningTime="2025-11-25 14:07:15.413176955 +0000 UTC m=+1244.575827405" Nov 25 14:07:18 crc kubenswrapper[4702]: I1125 14:07:18.423256 4702 generic.go:334] "Generic (PLEG): container finished" podID="6cf87935-5546-46f6-b8bb-f254e611faa7" containerID="2af2837b46377b1533459d4a0613c8e64b3a32f54338f03ae72128307b90e344" exitCode=0 Nov 25 14:07:18 crc kubenswrapper[4702]: I1125 14:07:18.423356 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-ldqjq" event={"ID":"6cf87935-5546-46f6-b8bb-f254e611faa7","Type":"ContainerDied","Data":"2af2837b46377b1533459d4a0613c8e64b3a32f54338f03ae72128307b90e344"} Nov 25 14:07:19 crc kubenswrapper[4702]: I1125 14:07:19.746977 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-ldqjq" Nov 25 14:07:19 crc kubenswrapper[4702]: I1125 14:07:19.780895 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6cf87935-5546-46f6-b8bb-f254e611faa7-db-sync-config-data\") pod \"6cf87935-5546-46f6-b8bb-f254e611faa7\" (UID: \"6cf87935-5546-46f6-b8bb-f254e611faa7\") " Nov 25 14:07:19 crc kubenswrapper[4702]: I1125 14:07:19.781026 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cf87935-5546-46f6-b8bb-f254e611faa7-combined-ca-bundle\") pod \"6cf87935-5546-46f6-b8bb-f254e611faa7\" (UID: \"6cf87935-5546-46f6-b8bb-f254e611faa7\") " Nov 25 14:07:19 crc kubenswrapper[4702]: I1125 14:07:19.781073 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nwwnq\" (UniqueName: \"kubernetes.io/projected/6cf87935-5546-46f6-b8bb-f254e611faa7-kube-api-access-nwwnq\") pod \"6cf87935-5546-46f6-b8bb-f254e611faa7\" (UID: \"6cf87935-5546-46f6-b8bb-f254e611faa7\") " Nov 25 14:07:19 crc kubenswrapper[4702]: I1125 14:07:19.781116 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cf87935-5546-46f6-b8bb-f254e611faa7-config-data\") pod \"6cf87935-5546-46f6-b8bb-f254e611faa7\" (UID: \"6cf87935-5546-46f6-b8bb-f254e611faa7\") " Nov 25 14:07:19 crc kubenswrapper[4702]: I1125 14:07:19.787535 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cf87935-5546-46f6-b8bb-f254e611faa7-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "6cf87935-5546-46f6-b8bb-f254e611faa7" (UID: "6cf87935-5546-46f6-b8bb-f254e611faa7"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:07:19 crc kubenswrapper[4702]: I1125 14:07:19.788836 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cf87935-5546-46f6-b8bb-f254e611faa7-kube-api-access-nwwnq" (OuterVolumeSpecName: "kube-api-access-nwwnq") pod "6cf87935-5546-46f6-b8bb-f254e611faa7" (UID: "6cf87935-5546-46f6-b8bb-f254e611faa7"). InnerVolumeSpecName "kube-api-access-nwwnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:07:19 crc kubenswrapper[4702]: I1125 14:07:19.805030 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cf87935-5546-46f6-b8bb-f254e611faa7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6cf87935-5546-46f6-b8bb-f254e611faa7" (UID: "6cf87935-5546-46f6-b8bb-f254e611faa7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:07:19 crc kubenswrapper[4702]: I1125 14:07:19.827689 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cf87935-5546-46f6-b8bb-f254e611faa7-config-data" (OuterVolumeSpecName: "config-data") pod "6cf87935-5546-46f6-b8bb-f254e611faa7" (UID: "6cf87935-5546-46f6-b8bb-f254e611faa7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:07:19 crc kubenswrapper[4702]: I1125 14:07:19.884547 4702 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cf87935-5546-46f6-b8bb-f254e611faa7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:19 crc kubenswrapper[4702]: I1125 14:07:19.884739 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nwwnq\" (UniqueName: \"kubernetes.io/projected/6cf87935-5546-46f6-b8bb-f254e611faa7-kube-api-access-nwwnq\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:19 crc kubenswrapper[4702]: I1125 14:07:19.884795 4702 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cf87935-5546-46f6-b8bb-f254e611faa7-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:19 crc kubenswrapper[4702]: I1125 14:07:19.884817 4702 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6cf87935-5546-46f6-b8bb-f254e611faa7-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:20 crc kubenswrapper[4702]: I1125 14:07:20.443850 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-ldqjq" event={"ID":"6cf87935-5546-46f6-b8bb-f254e611faa7","Type":"ContainerDied","Data":"69c8b54d60f1392c6f766827bd599cd038ec067a45741ad2f09fb2c727b687fe"} Nov 25 14:07:20 crc kubenswrapper[4702]: I1125 14:07:20.443905 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="69c8b54d60f1392c6f766827bd599cd038ec067a45741ad2f09fb2c727b687fe" Nov 25 14:07:20 crc kubenswrapper[4702]: I1125 14:07:20.443944 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-ldqjq" Nov 25 14:07:21 crc kubenswrapper[4702]: I1125 14:07:21.737869 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 14:07:21 crc kubenswrapper[4702]: E1125 14:07:21.738742 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cf87935-5546-46f6-b8bb-f254e611faa7" containerName="glance-db-sync" Nov 25 14:07:21 crc kubenswrapper[4702]: I1125 14:07:21.738769 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cf87935-5546-46f6-b8bb-f254e611faa7" containerName="glance-db-sync" Nov 25 14:07:21 crc kubenswrapper[4702]: I1125 14:07:21.739064 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cf87935-5546-46f6-b8bb-f254e611faa7" containerName="glance-db-sync" Nov 25 14:07:21 crc kubenswrapper[4702]: I1125 14:07:21.740381 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:21 crc kubenswrapper[4702]: I1125 14:07:21.745210 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 25 14:07:21 crc kubenswrapper[4702]: I1125 14:07:21.745644 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Nov 25 14:07:21 crc kubenswrapper[4702]: I1125 14:07:21.745876 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-public-svc" Nov 25 14:07:21 crc kubenswrapper[4702]: I1125 14:07:21.746358 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Nov 25 14:07:21 crc kubenswrapper[4702]: I1125 14:07:21.746797 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-internal-svc" Nov 25 14:07:21 crc kubenswrapper[4702]: I1125 14:07:21.748011 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-kjx4l" Nov 25 14:07:21 crc kubenswrapper[4702]: I1125 14:07:21.804934 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 14:07:21 crc kubenswrapper[4702]: I1125 14:07:21.938561 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4bd8326-0b68-48f7-87af-fc22c806f6bb-config-data\") pod \"glance-default-single-0\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:21 crc kubenswrapper[4702]: I1125 14:07:21.938667 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4bd8326-0b68-48f7-87af-fc22c806f6bb-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:21 crc kubenswrapper[4702]: I1125 14:07:21.938692 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4bd8326-0b68-48f7-87af-fc22c806f6bb-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:21 crc kubenswrapper[4702]: I1125 14:07:21.938715 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4bd8326-0b68-48f7-87af-fc22c806f6bb-logs\") pod \"glance-default-single-0\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:21 crc kubenswrapper[4702]: I1125 14:07:21.938847 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e4bd8326-0b68-48f7-87af-fc22c806f6bb-httpd-run\") pod \"glance-default-single-0\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:21 crc kubenswrapper[4702]: I1125 14:07:21.939286 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4bd8326-0b68-48f7-87af-fc22c806f6bb-scripts\") pod \"glance-default-single-0\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:21 crc kubenswrapper[4702]: I1125 14:07:21.939355 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7m5dt\" (UniqueName: \"kubernetes.io/projected/e4bd8326-0b68-48f7-87af-fc22c806f6bb-kube-api-access-7m5dt\") pod \"glance-default-single-0\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:21 crc kubenswrapper[4702]: I1125 14:07:21.939421 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-0\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:21 crc kubenswrapper[4702]: I1125 14:07:21.940117 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4bd8326-0b68-48f7-87af-fc22c806f6bb-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:22 crc kubenswrapper[4702]: I1125 14:07:22.041649 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4bd8326-0b68-48f7-87af-fc22c806f6bb-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:22 crc kubenswrapper[4702]: I1125 14:07:22.041726 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4bd8326-0b68-48f7-87af-fc22c806f6bb-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:22 crc kubenswrapper[4702]: I1125 14:07:22.041756 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4bd8326-0b68-48f7-87af-fc22c806f6bb-logs\") pod \"glance-default-single-0\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:22 crc kubenswrapper[4702]: I1125 14:07:22.041803 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e4bd8326-0b68-48f7-87af-fc22c806f6bb-httpd-run\") pod \"glance-default-single-0\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:22 crc kubenswrapper[4702]: I1125 14:07:22.041914 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4bd8326-0b68-48f7-87af-fc22c806f6bb-scripts\") pod \"glance-default-single-0\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:22 crc kubenswrapper[4702]: I1125 14:07:22.041947 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7m5dt\" (UniqueName: \"kubernetes.io/projected/e4bd8326-0b68-48f7-87af-fc22c806f6bb-kube-api-access-7m5dt\") pod \"glance-default-single-0\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:22 crc kubenswrapper[4702]: I1125 14:07:22.041985 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-0\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:22 crc kubenswrapper[4702]: I1125 14:07:22.042020 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4bd8326-0b68-48f7-87af-fc22c806f6bb-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:22 crc kubenswrapper[4702]: I1125 14:07:22.042490 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4bd8326-0b68-48f7-87af-fc22c806f6bb-logs\") pod \"glance-default-single-0\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:22 crc kubenswrapper[4702]: I1125 14:07:22.042491 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4bd8326-0b68-48f7-87af-fc22c806f6bb-config-data\") pod \"glance-default-single-0\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:22 crc kubenswrapper[4702]: I1125 14:07:22.043105 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e4bd8326-0b68-48f7-87af-fc22c806f6bb-httpd-run\") pod \"glance-default-single-0\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:22 crc kubenswrapper[4702]: I1125 14:07:22.043226 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-0\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:22 crc kubenswrapper[4702]: I1125 14:07:22.050053 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4bd8326-0b68-48f7-87af-fc22c806f6bb-scripts\") pod \"glance-default-single-0\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:22 crc kubenswrapper[4702]: I1125 14:07:22.050224 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4bd8326-0b68-48f7-87af-fc22c806f6bb-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:22 crc kubenswrapper[4702]: I1125 14:07:22.055141 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4bd8326-0b68-48f7-87af-fc22c806f6bb-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:22 crc kubenswrapper[4702]: I1125 14:07:22.056454 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4bd8326-0b68-48f7-87af-fc22c806f6bb-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:22 crc kubenswrapper[4702]: I1125 14:07:22.058891 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4bd8326-0b68-48f7-87af-fc22c806f6bb-config-data\") pod \"glance-default-single-0\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:22 crc kubenswrapper[4702]: I1125 14:07:22.068154 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7m5dt\" (UniqueName: \"kubernetes.io/projected/e4bd8326-0b68-48f7-87af-fc22c806f6bb-kube-api-access-7m5dt\") pod \"glance-default-single-0\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:22 crc kubenswrapper[4702]: I1125 14:07:22.080768 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-0\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:22 crc kubenswrapper[4702]: I1125 14:07:22.173256 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 14:07:22 crc kubenswrapper[4702]: I1125 14:07:22.174009 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:22 crc kubenswrapper[4702]: I1125 14:07:22.476681 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 14:07:23 crc kubenswrapper[4702]: I1125 14:07:23.475446 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"e4bd8326-0b68-48f7-87af-fc22c806f6bb","Type":"ContainerStarted","Data":"757d256ab21f4ed8de8976514c8363ee0849fae8c6e56d9d0262d226458cf3c9"} Nov 25 14:07:23 crc kubenswrapper[4702]: I1125 14:07:23.475786 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"e4bd8326-0b68-48f7-87af-fc22c806f6bb","Type":"ContainerStarted","Data":"5940ab4bb20a227f41efe56c3dcee277c0eb5cde07c8bdb366664ddeccb78ef4"} Nov 25 14:07:24 crc kubenswrapper[4702]: I1125 14:07:24.487840 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"e4bd8326-0b68-48f7-87af-fc22c806f6bb","Type":"ContainerStarted","Data":"75df23a67813aea6443b9dc7c13f9c188405f163a3524f9abbed209e2e930bc8"} Nov 25 14:07:24 crc kubenswrapper[4702]: I1125 14:07:24.488067 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="e4bd8326-0b68-48f7-87af-fc22c806f6bb" containerName="glance-log" containerID="cri-o://757d256ab21f4ed8de8976514c8363ee0849fae8c6e56d9d0262d226458cf3c9" gracePeriod=30 Nov 25 14:07:24 crc kubenswrapper[4702]: I1125 14:07:24.488156 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="e4bd8326-0b68-48f7-87af-fc22c806f6bb" containerName="glance-httpd" containerID="cri-o://75df23a67813aea6443b9dc7c13f9c188405f163a3524f9abbed209e2e930bc8" gracePeriod=30 Nov 25 14:07:24 crc kubenswrapper[4702]: I1125 14:07:24.525107 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=4.525070595 podStartE2EDuration="4.525070595s" podCreationTimestamp="2025-11-25 14:07:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 14:07:24.516916533 +0000 UTC m=+1253.679566983" watchObservedRunningTime="2025-11-25 14:07:24.525070595 +0000 UTC m=+1253.687721055" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.047916 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.202875 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4bd8326-0b68-48f7-87af-fc22c806f6bb-combined-ca-bundle\") pod \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.202967 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e4bd8326-0b68-48f7-87af-fc22c806f6bb-httpd-run\") pod \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.203010 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4bd8326-0b68-48f7-87af-fc22c806f6bb-scripts\") pod \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.203088 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.203115 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4bd8326-0b68-48f7-87af-fc22c806f6bb-logs\") pod \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.203138 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7m5dt\" (UniqueName: \"kubernetes.io/projected/e4bd8326-0b68-48f7-87af-fc22c806f6bb-kube-api-access-7m5dt\") pod \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.203189 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4bd8326-0b68-48f7-87af-fc22c806f6bb-internal-tls-certs\") pod \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.203252 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4bd8326-0b68-48f7-87af-fc22c806f6bb-config-data\") pod \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.203287 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4bd8326-0b68-48f7-87af-fc22c806f6bb-public-tls-certs\") pod \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\" (UID: \"e4bd8326-0b68-48f7-87af-fc22c806f6bb\") " Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.204153 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4bd8326-0b68-48f7-87af-fc22c806f6bb-logs" (OuterVolumeSpecName: "logs") pod "e4bd8326-0b68-48f7-87af-fc22c806f6bb" (UID: "e4bd8326-0b68-48f7-87af-fc22c806f6bb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.204388 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4bd8326-0b68-48f7-87af-fc22c806f6bb-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e4bd8326-0b68-48f7-87af-fc22c806f6bb" (UID: "e4bd8326-0b68-48f7-87af-fc22c806f6bb"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.212129 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4bd8326-0b68-48f7-87af-fc22c806f6bb-scripts" (OuterVolumeSpecName: "scripts") pod "e4bd8326-0b68-48f7-87af-fc22c806f6bb" (UID: "e4bd8326-0b68-48f7-87af-fc22c806f6bb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.212693 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4bd8326-0b68-48f7-87af-fc22c806f6bb-kube-api-access-7m5dt" (OuterVolumeSpecName: "kube-api-access-7m5dt") pod "e4bd8326-0b68-48f7-87af-fc22c806f6bb" (UID: "e4bd8326-0b68-48f7-87af-fc22c806f6bb"). InnerVolumeSpecName "kube-api-access-7m5dt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.215846 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "e4bd8326-0b68-48f7-87af-fc22c806f6bb" (UID: "e4bd8326-0b68-48f7-87af-fc22c806f6bb"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.232077 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4bd8326-0b68-48f7-87af-fc22c806f6bb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e4bd8326-0b68-48f7-87af-fc22c806f6bb" (UID: "e4bd8326-0b68-48f7-87af-fc22c806f6bb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.250612 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4bd8326-0b68-48f7-87af-fc22c806f6bb-config-data" (OuterVolumeSpecName: "config-data") pod "e4bd8326-0b68-48f7-87af-fc22c806f6bb" (UID: "e4bd8326-0b68-48f7-87af-fc22c806f6bb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.252205 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4bd8326-0b68-48f7-87af-fc22c806f6bb-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e4bd8326-0b68-48f7-87af-fc22c806f6bb" (UID: "e4bd8326-0b68-48f7-87af-fc22c806f6bb"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.257411 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4bd8326-0b68-48f7-87af-fc22c806f6bb-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e4bd8326-0b68-48f7-87af-fc22c806f6bb" (UID: "e4bd8326-0b68-48f7-87af-fc22c806f6bb"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.305729 4702 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4bd8326-0b68-48f7-87af-fc22c806f6bb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.305776 4702 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e4bd8326-0b68-48f7-87af-fc22c806f6bb-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.305788 4702 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4bd8326-0b68-48f7-87af-fc22c806f6bb-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.305837 4702 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.305849 4702 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4bd8326-0b68-48f7-87af-fc22c806f6bb-logs\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.305869 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7m5dt\" (UniqueName: \"kubernetes.io/projected/e4bd8326-0b68-48f7-87af-fc22c806f6bb-kube-api-access-7m5dt\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.305885 4702 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4bd8326-0b68-48f7-87af-fc22c806f6bb-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.305895 4702 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4bd8326-0b68-48f7-87af-fc22c806f6bb-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.305906 4702 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4bd8326-0b68-48f7-87af-fc22c806f6bb-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.324844 4702 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.407038 4702 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.502211 4702 generic.go:334] "Generic (PLEG): container finished" podID="e4bd8326-0b68-48f7-87af-fc22c806f6bb" containerID="75df23a67813aea6443b9dc7c13f9c188405f163a3524f9abbed209e2e930bc8" exitCode=0 Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.502307 4702 generic.go:334] "Generic (PLEG): container finished" podID="e4bd8326-0b68-48f7-87af-fc22c806f6bb" containerID="757d256ab21f4ed8de8976514c8363ee0849fae8c6e56d9d0262d226458cf3c9" exitCode=143 Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.502343 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"e4bd8326-0b68-48f7-87af-fc22c806f6bb","Type":"ContainerDied","Data":"75df23a67813aea6443b9dc7c13f9c188405f163a3524f9abbed209e2e930bc8"} Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.502364 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.502407 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"e4bd8326-0b68-48f7-87af-fc22c806f6bb","Type":"ContainerDied","Data":"757d256ab21f4ed8de8976514c8363ee0849fae8c6e56d9d0262d226458cf3c9"} Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.502431 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"e4bd8326-0b68-48f7-87af-fc22c806f6bb","Type":"ContainerDied","Data":"5940ab4bb20a227f41efe56c3dcee277c0eb5cde07c8bdb366664ddeccb78ef4"} Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.502459 4702 scope.go:117] "RemoveContainer" containerID="75df23a67813aea6443b9dc7c13f9c188405f163a3524f9abbed209e2e930bc8" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.528428 4702 scope.go:117] "RemoveContainer" containerID="757d256ab21f4ed8de8976514c8363ee0849fae8c6e56d9d0262d226458cf3c9" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.546888 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.565482 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.575227 4702 scope.go:117] "RemoveContainer" containerID="75df23a67813aea6443b9dc7c13f9c188405f163a3524f9abbed209e2e930bc8" Nov 25 14:07:25 crc kubenswrapper[4702]: E1125 14:07:25.577046 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75df23a67813aea6443b9dc7c13f9c188405f163a3524f9abbed209e2e930bc8\": container with ID starting with 75df23a67813aea6443b9dc7c13f9c188405f163a3524f9abbed209e2e930bc8 not found: ID does not exist" containerID="75df23a67813aea6443b9dc7c13f9c188405f163a3524f9abbed209e2e930bc8" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.577086 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75df23a67813aea6443b9dc7c13f9c188405f163a3524f9abbed209e2e930bc8"} err="failed to get container status \"75df23a67813aea6443b9dc7c13f9c188405f163a3524f9abbed209e2e930bc8\": rpc error: code = NotFound desc = could not find container \"75df23a67813aea6443b9dc7c13f9c188405f163a3524f9abbed209e2e930bc8\": container with ID starting with 75df23a67813aea6443b9dc7c13f9c188405f163a3524f9abbed209e2e930bc8 not found: ID does not exist" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.577140 4702 scope.go:117] "RemoveContainer" containerID="757d256ab21f4ed8de8976514c8363ee0849fae8c6e56d9d0262d226458cf3c9" Nov 25 14:07:25 crc kubenswrapper[4702]: E1125 14:07:25.577691 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"757d256ab21f4ed8de8976514c8363ee0849fae8c6e56d9d0262d226458cf3c9\": container with ID starting with 757d256ab21f4ed8de8976514c8363ee0849fae8c6e56d9d0262d226458cf3c9 not found: ID does not exist" containerID="757d256ab21f4ed8de8976514c8363ee0849fae8c6e56d9d0262d226458cf3c9" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.577765 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"757d256ab21f4ed8de8976514c8363ee0849fae8c6e56d9d0262d226458cf3c9"} err="failed to get container status \"757d256ab21f4ed8de8976514c8363ee0849fae8c6e56d9d0262d226458cf3c9\": rpc error: code = NotFound desc = could not find container \"757d256ab21f4ed8de8976514c8363ee0849fae8c6e56d9d0262d226458cf3c9\": container with ID starting with 757d256ab21f4ed8de8976514c8363ee0849fae8c6e56d9d0262d226458cf3c9 not found: ID does not exist" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.577787 4702 scope.go:117] "RemoveContainer" containerID="75df23a67813aea6443b9dc7c13f9c188405f163a3524f9abbed209e2e930bc8" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.578442 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75df23a67813aea6443b9dc7c13f9c188405f163a3524f9abbed209e2e930bc8"} err="failed to get container status \"75df23a67813aea6443b9dc7c13f9c188405f163a3524f9abbed209e2e930bc8\": rpc error: code = NotFound desc = could not find container \"75df23a67813aea6443b9dc7c13f9c188405f163a3524f9abbed209e2e930bc8\": container with ID starting with 75df23a67813aea6443b9dc7c13f9c188405f163a3524f9abbed209e2e930bc8 not found: ID does not exist" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.578695 4702 scope.go:117] "RemoveContainer" containerID="757d256ab21f4ed8de8976514c8363ee0849fae8c6e56d9d0262d226458cf3c9" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.578767 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 14:07:25 crc kubenswrapper[4702]: E1125 14:07:25.579708 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4bd8326-0b68-48f7-87af-fc22c806f6bb" containerName="glance-httpd" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.579735 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4bd8326-0b68-48f7-87af-fc22c806f6bb" containerName="glance-httpd" Nov 25 14:07:25 crc kubenswrapper[4702]: E1125 14:07:25.580034 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4bd8326-0b68-48f7-87af-fc22c806f6bb" containerName="glance-log" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.580045 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4bd8326-0b68-48f7-87af-fc22c806f6bb" containerName="glance-log" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.579925 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"757d256ab21f4ed8de8976514c8363ee0849fae8c6e56d9d0262d226458cf3c9"} err="failed to get container status \"757d256ab21f4ed8de8976514c8363ee0849fae8c6e56d9d0262d226458cf3c9\": rpc error: code = NotFound desc = could not find container \"757d256ab21f4ed8de8976514c8363ee0849fae8c6e56d9d0262d226458cf3c9\": container with ID starting with 757d256ab21f4ed8de8976514c8363ee0849fae8c6e56d9d0262d226458cf3c9 not found: ID does not exist" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.580267 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4bd8326-0b68-48f7-87af-fc22c806f6bb" containerName="glance-httpd" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.580292 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4bd8326-0b68-48f7-87af-fc22c806f6bb" containerName="glance-log" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.581157 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.586042 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.586625 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-public-svc" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.587248 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.587322 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-internal-svc" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.587761 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.587986 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-kjx4l" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.601682 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.712358 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.712458 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74a328c5-c8cd-49e6-ad80-de2b34117428-httpd-run\") pod \"glance-default-single-0\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.712496 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-scripts\") pod \"glance-default-single-0\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.712519 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74a328c5-c8cd-49e6-ad80-de2b34117428-logs\") pod \"glance-default-single-0\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.712551 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.712571 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-0\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.712767 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-config-data\") pod \"glance-default-single-0\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.712988 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mjnq\" (UniqueName: \"kubernetes.io/projected/74a328c5-c8cd-49e6-ad80-de2b34117428-kube-api-access-4mjnq\") pod \"glance-default-single-0\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.713146 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.815141 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-config-data\") pod \"glance-default-single-0\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.815302 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mjnq\" (UniqueName: \"kubernetes.io/projected/74a328c5-c8cd-49e6-ad80-de2b34117428-kube-api-access-4mjnq\") pod \"glance-default-single-0\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.815369 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.815451 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.815522 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74a328c5-c8cd-49e6-ad80-de2b34117428-httpd-run\") pod \"glance-default-single-0\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.815567 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-scripts\") pod \"glance-default-single-0\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.815602 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74a328c5-c8cd-49e6-ad80-de2b34117428-logs\") pod \"glance-default-single-0\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.815654 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-0\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.815688 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.817135 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-0\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.817431 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74a328c5-c8cd-49e6-ad80-de2b34117428-logs\") pod \"glance-default-single-0\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.817515 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74a328c5-c8cd-49e6-ad80-de2b34117428-httpd-run\") pod \"glance-default-single-0\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.821827 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4bd8326-0b68-48f7-87af-fc22c806f6bb" path="/var/lib/kubelet/pods/e4bd8326-0b68-48f7-87af-fc22c806f6bb/volumes" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.825290 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.825608 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.826420 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-scripts\") pod \"glance-default-single-0\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.826880 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-config-data\") pod \"glance-default-single-0\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.830865 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.840881 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mjnq\" (UniqueName: \"kubernetes.io/projected/74a328c5-c8cd-49e6-ad80-de2b34117428-kube-api-access-4mjnq\") pod \"glance-default-single-0\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.848997 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-0\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:25 crc kubenswrapper[4702]: I1125 14:07:25.935333 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:26 crc kubenswrapper[4702]: I1125 14:07:26.386439 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 14:07:26 crc kubenswrapper[4702]: W1125 14:07:26.388066 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74a328c5_c8cd_49e6_ad80_de2b34117428.slice/crio-0494141c668f009018d37e2e59c664d87f96a24249d60fb74c389b4fa6abd9de WatchSource:0}: Error finding container 0494141c668f009018d37e2e59c664d87f96a24249d60fb74c389b4fa6abd9de: Status 404 returned error can't find the container with id 0494141c668f009018d37e2e59c664d87f96a24249d60fb74c389b4fa6abd9de Nov 25 14:07:26 crc kubenswrapper[4702]: I1125 14:07:26.519657 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"74a328c5-c8cd-49e6-ad80-de2b34117428","Type":"ContainerStarted","Data":"0494141c668f009018d37e2e59c664d87f96a24249d60fb74c389b4fa6abd9de"} Nov 25 14:07:27 crc kubenswrapper[4702]: I1125 14:07:27.532314 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"74a328c5-c8cd-49e6-ad80-de2b34117428","Type":"ContainerStarted","Data":"621cc723417ff124d718a266eac42bad0bd5ad408d46139392b336444b8934ee"} Nov 25 14:07:28 crc kubenswrapper[4702]: I1125 14:07:28.542862 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"74a328c5-c8cd-49e6-ad80-de2b34117428","Type":"ContainerStarted","Data":"3790fca52560d67bfdd3de6e066c1b60b288208f5af16b96fadd1b3c83fc6e0a"} Nov 25 14:07:28 crc kubenswrapper[4702]: I1125 14:07:28.574481 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=3.574454178 podStartE2EDuration="3.574454178s" podCreationTimestamp="2025-11-25 14:07:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 14:07:28.567726612 +0000 UTC m=+1257.730377082" watchObservedRunningTime="2025-11-25 14:07:28.574454178 +0000 UTC m=+1257.737104628" Nov 25 14:07:33 crc kubenswrapper[4702]: I1125 14:07:33.962427 4702 patch_prober.go:28] interesting pod/machine-config-daemon-qqr76 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 14:07:33 crc kubenswrapper[4702]: I1125 14:07:33.963191 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 14:07:33 crc kubenswrapper[4702]: I1125 14:07:33.963277 4702 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" Nov 25 14:07:33 crc kubenswrapper[4702]: I1125 14:07:33.964071 4702 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"de4c184eb16dc995f1ebd5c48e0561cca0a21907587a6c388b21643c8a15afa4"} pod="openshift-machine-config-operator/machine-config-daemon-qqr76" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 14:07:33 crc kubenswrapper[4702]: I1125 14:07:33.964129 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" containerID="cri-o://de4c184eb16dc995f1ebd5c48e0561cca0a21907587a6c388b21643c8a15afa4" gracePeriod=600 Nov 25 14:07:34 crc kubenswrapper[4702]: I1125 14:07:34.595418 4702 generic.go:334] "Generic (PLEG): container finished" podID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerID="de4c184eb16dc995f1ebd5c48e0561cca0a21907587a6c388b21643c8a15afa4" exitCode=0 Nov 25 14:07:34 crc kubenswrapper[4702]: I1125 14:07:34.595491 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" event={"ID":"3c491818-b0bb-4d82-a031-96e2dfac8c27","Type":"ContainerDied","Data":"de4c184eb16dc995f1ebd5c48e0561cca0a21907587a6c388b21643c8a15afa4"} Nov 25 14:07:34 crc kubenswrapper[4702]: I1125 14:07:34.595997 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" event={"ID":"3c491818-b0bb-4d82-a031-96e2dfac8c27","Type":"ContainerStarted","Data":"3c4335b16016506091cd0c40c0eeaec31932b0b8e710fd86970dfc2371661ab2"} Nov 25 14:07:34 crc kubenswrapper[4702]: I1125 14:07:34.596024 4702 scope.go:117] "RemoveContainer" containerID="03504dcb763332c521878c39bb07b19a232e8fc711f4c606ea5726c5c5f4da1a" Nov 25 14:07:35 crc kubenswrapper[4702]: I1125 14:07:35.936453 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:35 crc kubenswrapper[4702]: I1125 14:07:35.936989 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:35 crc kubenswrapper[4702]: I1125 14:07:35.964474 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:35 crc kubenswrapper[4702]: I1125 14:07:35.974740 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:36 crc kubenswrapper[4702]: I1125 14:07:36.615395 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:36 crc kubenswrapper[4702]: I1125 14:07:36.615469 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:38 crc kubenswrapper[4702]: I1125 14:07:38.774031 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:38 crc kubenswrapper[4702]: I1125 14:07:38.774753 4702 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 14:07:38 crc kubenswrapper[4702]: I1125 14:07:38.775923 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:39 crc kubenswrapper[4702]: I1125 14:07:39.731931 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-ldqjq"] Nov 25 14:07:39 crc kubenswrapper[4702]: I1125 14:07:39.737303 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-ldqjq"] Nov 25 14:07:39 crc kubenswrapper[4702]: E1125 14:07:39.758377 4702 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Nov 25 14:07:39 crc kubenswrapper[4702]: E1125 14:07:39.758490 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-scripts podName:74a328c5-c8cd-49e6-ad80-de2b34117428 nodeName:}" failed. No retries permitted until 2025-11-25 14:07:40.258463596 +0000 UTC m=+1269.421114046 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-scripts") pod "glance-default-single-0" (UID: "74a328c5-c8cd-49e6-ad80-de2b34117428") : secret "glance-scripts" not found Nov 25 14:07:39 crc kubenswrapper[4702]: I1125 14:07:39.869530 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cf87935-5546-46f6-b8bb-f254e611faa7" path="/var/lib/kubelet/pods/6cf87935-5546-46f6-b8bb-f254e611faa7/volumes" Nov 25 14:07:39 crc kubenswrapper[4702]: I1125 14:07:39.876293 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 14:07:39 crc kubenswrapper[4702]: I1125 14:07:39.891580 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glancefaf0-account-delete-lm4vl"] Nov 25 14:07:39 crc kubenswrapper[4702]: I1125 14:07:39.893443 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancefaf0-account-delete-lm4vl" Nov 25 14:07:39 crc kubenswrapper[4702]: I1125 14:07:39.908555 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glancefaf0-account-delete-lm4vl"] Nov 25 14:07:39 crc kubenswrapper[4702]: I1125 14:07:39.971501 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8xzr\" (UniqueName: \"kubernetes.io/projected/617d6aec-6354-4119-b0bd-48f71ba9b336-kube-api-access-h8xzr\") pod \"glancefaf0-account-delete-lm4vl\" (UID: \"617d6aec-6354-4119-b0bd-48f71ba9b336\") " pod="glance-kuttl-tests/glancefaf0-account-delete-lm4vl" Nov 25 14:07:39 crc kubenswrapper[4702]: I1125 14:07:39.971925 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/617d6aec-6354-4119-b0bd-48f71ba9b336-operator-scripts\") pod \"glancefaf0-account-delete-lm4vl\" (UID: \"617d6aec-6354-4119-b0bd-48f71ba9b336\") " pod="glance-kuttl-tests/glancefaf0-account-delete-lm4vl" Nov 25 14:07:40 crc kubenswrapper[4702]: I1125 14:07:40.073224 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8xzr\" (UniqueName: \"kubernetes.io/projected/617d6aec-6354-4119-b0bd-48f71ba9b336-kube-api-access-h8xzr\") pod \"glancefaf0-account-delete-lm4vl\" (UID: \"617d6aec-6354-4119-b0bd-48f71ba9b336\") " pod="glance-kuttl-tests/glancefaf0-account-delete-lm4vl" Nov 25 14:07:40 crc kubenswrapper[4702]: I1125 14:07:40.073671 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/617d6aec-6354-4119-b0bd-48f71ba9b336-operator-scripts\") pod \"glancefaf0-account-delete-lm4vl\" (UID: \"617d6aec-6354-4119-b0bd-48f71ba9b336\") " pod="glance-kuttl-tests/glancefaf0-account-delete-lm4vl" Nov 25 14:07:40 crc kubenswrapper[4702]: I1125 14:07:40.074485 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/617d6aec-6354-4119-b0bd-48f71ba9b336-operator-scripts\") pod \"glancefaf0-account-delete-lm4vl\" (UID: \"617d6aec-6354-4119-b0bd-48f71ba9b336\") " pod="glance-kuttl-tests/glancefaf0-account-delete-lm4vl" Nov 25 14:07:40 crc kubenswrapper[4702]: I1125 14:07:40.094540 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8xzr\" (UniqueName: \"kubernetes.io/projected/617d6aec-6354-4119-b0bd-48f71ba9b336-kube-api-access-h8xzr\") pod \"glancefaf0-account-delete-lm4vl\" (UID: \"617d6aec-6354-4119-b0bd-48f71ba9b336\") " pod="glance-kuttl-tests/glancefaf0-account-delete-lm4vl" Nov 25 14:07:40 crc kubenswrapper[4702]: I1125 14:07:40.226305 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancefaf0-account-delete-lm4vl" Nov 25 14:07:40 crc kubenswrapper[4702]: E1125 14:07:40.278899 4702 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Nov 25 14:07:40 crc kubenswrapper[4702]: E1125 14:07:40.278987 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-scripts podName:74a328c5-c8cd-49e6-ad80-de2b34117428 nodeName:}" failed. No retries permitted until 2025-11-25 14:07:41.27895425 +0000 UTC m=+1270.441604700 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-scripts") pod "glance-default-single-0" (UID: "74a328c5-c8cd-49e6-ad80-de2b34117428") : secret "glance-scripts" not found Nov 25 14:07:40 crc kubenswrapper[4702]: I1125 14:07:40.653008 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="74a328c5-c8cd-49e6-ad80-de2b34117428" containerName="glance-log" containerID="cri-o://621cc723417ff124d718a266eac42bad0bd5ad408d46139392b336444b8934ee" gracePeriod=30 Nov 25 14:07:40 crc kubenswrapper[4702]: I1125 14:07:40.653400 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="74a328c5-c8cd-49e6-ad80-de2b34117428" containerName="glance-httpd" containerID="cri-o://3790fca52560d67bfdd3de6e066c1b60b288208f5af16b96fadd1b3c83fc6e0a" gracePeriod=30 Nov 25 14:07:40 crc kubenswrapper[4702]: I1125 14:07:40.654172 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glancefaf0-account-delete-lm4vl"] Nov 25 14:07:41 crc kubenswrapper[4702]: E1125 14:07:41.295112 4702 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Nov 25 14:07:41 crc kubenswrapper[4702]: E1125 14:07:41.295225 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-scripts podName:74a328c5-c8cd-49e6-ad80-de2b34117428 nodeName:}" failed. No retries permitted until 2025-11-25 14:07:43.295196989 +0000 UTC m=+1272.457847439 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-scripts") pod "glance-default-single-0" (UID: "74a328c5-c8cd-49e6-ad80-de2b34117428") : secret "glance-scripts" not found Nov 25 14:07:41 crc kubenswrapper[4702]: I1125 14:07:41.665086 4702 generic.go:334] "Generic (PLEG): container finished" podID="74a328c5-c8cd-49e6-ad80-de2b34117428" containerID="621cc723417ff124d718a266eac42bad0bd5ad408d46139392b336444b8934ee" exitCode=143 Nov 25 14:07:41 crc kubenswrapper[4702]: I1125 14:07:41.665171 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"74a328c5-c8cd-49e6-ad80-de2b34117428","Type":"ContainerDied","Data":"621cc723417ff124d718a266eac42bad0bd5ad408d46139392b336444b8934ee"} Nov 25 14:07:41 crc kubenswrapper[4702]: I1125 14:07:41.667973 4702 generic.go:334] "Generic (PLEG): container finished" podID="617d6aec-6354-4119-b0bd-48f71ba9b336" containerID="bcfe02a179b6890c70d2d8c7f50bc272681575257f235aee7da8e4e3d0c57fe6" exitCode=0 Nov 25 14:07:41 crc kubenswrapper[4702]: I1125 14:07:41.668004 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancefaf0-account-delete-lm4vl" event={"ID":"617d6aec-6354-4119-b0bd-48f71ba9b336","Type":"ContainerDied","Data":"bcfe02a179b6890c70d2d8c7f50bc272681575257f235aee7da8e4e3d0c57fe6"} Nov 25 14:07:41 crc kubenswrapper[4702]: I1125 14:07:41.668022 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancefaf0-account-delete-lm4vl" event={"ID":"617d6aec-6354-4119-b0bd-48f71ba9b336","Type":"ContainerStarted","Data":"e0c19482e38ec8adf4622b5e2a92e1ee4bb74bb8c598f945f1b8ebd5a80d4625"} Nov 25 14:07:42 crc kubenswrapper[4702]: I1125 14:07:42.992950 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancefaf0-account-delete-lm4vl" Nov 25 14:07:43 crc kubenswrapper[4702]: I1125 14:07:43.022080 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/617d6aec-6354-4119-b0bd-48f71ba9b336-operator-scripts\") pod \"617d6aec-6354-4119-b0bd-48f71ba9b336\" (UID: \"617d6aec-6354-4119-b0bd-48f71ba9b336\") " Nov 25 14:07:43 crc kubenswrapper[4702]: I1125 14:07:43.022258 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8xzr\" (UniqueName: \"kubernetes.io/projected/617d6aec-6354-4119-b0bd-48f71ba9b336-kube-api-access-h8xzr\") pod \"617d6aec-6354-4119-b0bd-48f71ba9b336\" (UID: \"617d6aec-6354-4119-b0bd-48f71ba9b336\") " Nov 25 14:07:43 crc kubenswrapper[4702]: I1125 14:07:43.022770 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/617d6aec-6354-4119-b0bd-48f71ba9b336-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "617d6aec-6354-4119-b0bd-48f71ba9b336" (UID: "617d6aec-6354-4119-b0bd-48f71ba9b336"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 14:07:43 crc kubenswrapper[4702]: I1125 14:07:43.030467 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/617d6aec-6354-4119-b0bd-48f71ba9b336-kube-api-access-h8xzr" (OuterVolumeSpecName: "kube-api-access-h8xzr") pod "617d6aec-6354-4119-b0bd-48f71ba9b336" (UID: "617d6aec-6354-4119-b0bd-48f71ba9b336"). InnerVolumeSpecName "kube-api-access-h8xzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:07:43 crc kubenswrapper[4702]: I1125 14:07:43.124147 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8xzr\" (UniqueName: \"kubernetes.io/projected/617d6aec-6354-4119-b0bd-48f71ba9b336-kube-api-access-h8xzr\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:43 crc kubenswrapper[4702]: I1125 14:07:43.124185 4702 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/617d6aec-6354-4119-b0bd-48f71ba9b336-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:43 crc kubenswrapper[4702]: E1125 14:07:43.327200 4702 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Nov 25 14:07:43 crc kubenswrapper[4702]: E1125 14:07:43.327333 4702 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-scripts podName:74a328c5-c8cd-49e6-ad80-de2b34117428 nodeName:}" failed. No retries permitted until 2025-11-25 14:07:47.327309619 +0000 UTC m=+1276.489960089 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-scripts") pod "glance-default-single-0" (UID: "74a328c5-c8cd-49e6-ad80-de2b34117428") : secret "glance-scripts" not found Nov 25 14:07:43 crc kubenswrapper[4702]: I1125 14:07:43.687898 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancefaf0-account-delete-lm4vl" event={"ID":"617d6aec-6354-4119-b0bd-48f71ba9b336","Type":"ContainerDied","Data":"e0c19482e38ec8adf4622b5e2a92e1ee4bb74bb8c598f945f1b8ebd5a80d4625"} Nov 25 14:07:43 crc kubenswrapper[4702]: I1125 14:07:43.687967 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0c19482e38ec8adf4622b5e2a92e1ee4bb74bb8c598f945f1b8ebd5a80d4625" Nov 25 14:07:43 crc kubenswrapper[4702]: I1125 14:07:43.687934 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancefaf0-account-delete-lm4vl" Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.310430 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.452912 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-config-data\") pod \"74a328c5-c8cd-49e6-ad80-de2b34117428\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.452988 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-scripts\") pod \"74a328c5-c8cd-49e6-ad80-de2b34117428\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.453079 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74a328c5-c8cd-49e6-ad80-de2b34117428-logs\") pod \"74a328c5-c8cd-49e6-ad80-de2b34117428\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.453161 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-public-tls-certs\") pod \"74a328c5-c8cd-49e6-ad80-de2b34117428\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.453209 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"74a328c5-c8cd-49e6-ad80-de2b34117428\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.453252 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-internal-tls-certs\") pod \"74a328c5-c8cd-49e6-ad80-de2b34117428\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.453858 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74a328c5-c8cd-49e6-ad80-de2b34117428-httpd-run\") pod \"74a328c5-c8cd-49e6-ad80-de2b34117428\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.453901 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74a328c5-c8cd-49e6-ad80-de2b34117428-logs" (OuterVolumeSpecName: "logs") pod "74a328c5-c8cd-49e6-ad80-de2b34117428" (UID: "74a328c5-c8cd-49e6-ad80-de2b34117428"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.453923 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-combined-ca-bundle\") pod \"74a328c5-c8cd-49e6-ad80-de2b34117428\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.454145 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mjnq\" (UniqueName: \"kubernetes.io/projected/74a328c5-c8cd-49e6-ad80-de2b34117428-kube-api-access-4mjnq\") pod \"74a328c5-c8cd-49e6-ad80-de2b34117428\" (UID: \"74a328c5-c8cd-49e6-ad80-de2b34117428\") " Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.454597 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74a328c5-c8cd-49e6-ad80-de2b34117428-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "74a328c5-c8cd-49e6-ad80-de2b34117428" (UID: "74a328c5-c8cd-49e6-ad80-de2b34117428"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.455114 4702 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74a328c5-c8cd-49e6-ad80-de2b34117428-logs\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.455148 4702 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74a328c5-c8cd-49e6-ad80-de2b34117428-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.459771 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-scripts" (OuterVolumeSpecName: "scripts") pod "74a328c5-c8cd-49e6-ad80-de2b34117428" (UID: "74a328c5-c8cd-49e6-ad80-de2b34117428"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.475503 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "74a328c5-c8cd-49e6-ad80-de2b34117428" (UID: "74a328c5-c8cd-49e6-ad80-de2b34117428"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.475542 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74a328c5-c8cd-49e6-ad80-de2b34117428-kube-api-access-4mjnq" (OuterVolumeSpecName: "kube-api-access-4mjnq") pod "74a328c5-c8cd-49e6-ad80-de2b34117428" (UID: "74a328c5-c8cd-49e6-ad80-de2b34117428"). InnerVolumeSpecName "kube-api-access-4mjnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.480023 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "74a328c5-c8cd-49e6-ad80-de2b34117428" (UID: "74a328c5-c8cd-49e6-ad80-de2b34117428"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.498355 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-config-data" (OuterVolumeSpecName: "config-data") pod "74a328c5-c8cd-49e6-ad80-de2b34117428" (UID: "74a328c5-c8cd-49e6-ad80-de2b34117428"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.505474 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "74a328c5-c8cd-49e6-ad80-de2b34117428" (UID: "74a328c5-c8cd-49e6-ad80-de2b34117428"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.516829 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "74a328c5-c8cd-49e6-ad80-de2b34117428" (UID: "74a328c5-c8cd-49e6-ad80-de2b34117428"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.557033 4702 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.557116 4702 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.557129 4702 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.557141 4702 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.557151 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mjnq\" (UniqueName: \"kubernetes.io/projected/74a328c5-c8cd-49e6-ad80-de2b34117428-kube-api-access-4mjnq\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.557164 4702 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.557172 4702 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74a328c5-c8cd-49e6-ad80-de2b34117428-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.570704 4702 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.659092 4702 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.699377 4702 generic.go:334] "Generic (PLEG): container finished" podID="74a328c5-c8cd-49e6-ad80-de2b34117428" containerID="3790fca52560d67bfdd3de6e066c1b60b288208f5af16b96fadd1b3c83fc6e0a" exitCode=0 Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.699437 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"74a328c5-c8cd-49e6-ad80-de2b34117428","Type":"ContainerDied","Data":"3790fca52560d67bfdd3de6e066c1b60b288208f5af16b96fadd1b3c83fc6e0a"} Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.699472 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"74a328c5-c8cd-49e6-ad80-de2b34117428","Type":"ContainerDied","Data":"0494141c668f009018d37e2e59c664d87f96a24249d60fb74c389b4fa6abd9de"} Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.699495 4702 scope.go:117] "RemoveContainer" containerID="3790fca52560d67bfdd3de6e066c1b60b288208f5af16b96fadd1b3c83fc6e0a" Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.699639 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.736283 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.739898 4702 scope.go:117] "RemoveContainer" containerID="621cc723417ff124d718a266eac42bad0bd5ad408d46139392b336444b8934ee" Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.746726 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.789243 4702 scope.go:117] "RemoveContainer" containerID="3790fca52560d67bfdd3de6e066c1b60b288208f5af16b96fadd1b3c83fc6e0a" Nov 25 14:07:44 crc kubenswrapper[4702]: E1125 14:07:44.789944 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3790fca52560d67bfdd3de6e066c1b60b288208f5af16b96fadd1b3c83fc6e0a\": container with ID starting with 3790fca52560d67bfdd3de6e066c1b60b288208f5af16b96fadd1b3c83fc6e0a not found: ID does not exist" containerID="3790fca52560d67bfdd3de6e066c1b60b288208f5af16b96fadd1b3c83fc6e0a" Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.789987 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3790fca52560d67bfdd3de6e066c1b60b288208f5af16b96fadd1b3c83fc6e0a"} err="failed to get container status \"3790fca52560d67bfdd3de6e066c1b60b288208f5af16b96fadd1b3c83fc6e0a\": rpc error: code = NotFound desc = could not find container \"3790fca52560d67bfdd3de6e066c1b60b288208f5af16b96fadd1b3c83fc6e0a\": container with ID starting with 3790fca52560d67bfdd3de6e066c1b60b288208f5af16b96fadd1b3c83fc6e0a not found: ID does not exist" Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.790018 4702 scope.go:117] "RemoveContainer" containerID="621cc723417ff124d718a266eac42bad0bd5ad408d46139392b336444b8934ee" Nov 25 14:07:44 crc kubenswrapper[4702]: E1125 14:07:44.790429 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"621cc723417ff124d718a266eac42bad0bd5ad408d46139392b336444b8934ee\": container with ID starting with 621cc723417ff124d718a266eac42bad0bd5ad408d46139392b336444b8934ee not found: ID does not exist" containerID="621cc723417ff124d718a266eac42bad0bd5ad408d46139392b336444b8934ee" Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.790458 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"621cc723417ff124d718a266eac42bad0bd5ad408d46139392b336444b8934ee"} err="failed to get container status \"621cc723417ff124d718a266eac42bad0bd5ad408d46139392b336444b8934ee\": rpc error: code = NotFound desc = could not find container \"621cc723417ff124d718a266eac42bad0bd5ad408d46139392b336444b8934ee\": container with ID starting with 621cc723417ff124d718a266eac42bad0bd5ad408d46139392b336444b8934ee not found: ID does not exist" Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.838729 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-pth7k"] Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.847662 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-pth7k"] Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.873736 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-faf0-account-create-update-p4425"] Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.881187 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glancefaf0-account-delete-lm4vl"] Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.888296 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glancefaf0-account-delete-lm4vl"] Nov 25 14:07:44 crc kubenswrapper[4702]: I1125 14:07:44.893499 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-faf0-account-create-update-p4425"] Nov 25 14:07:45 crc kubenswrapper[4702]: I1125 14:07:45.774544 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-zbqwl"] Nov 25 14:07:45 crc kubenswrapper[4702]: E1125 14:07:45.776120 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74a328c5-c8cd-49e6-ad80-de2b34117428" containerName="glance-httpd" Nov 25 14:07:45 crc kubenswrapper[4702]: I1125 14:07:45.776204 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="74a328c5-c8cd-49e6-ad80-de2b34117428" containerName="glance-httpd" Nov 25 14:07:45 crc kubenswrapper[4702]: E1125 14:07:45.776288 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74a328c5-c8cd-49e6-ad80-de2b34117428" containerName="glance-log" Nov 25 14:07:45 crc kubenswrapper[4702]: I1125 14:07:45.776344 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="74a328c5-c8cd-49e6-ad80-de2b34117428" containerName="glance-log" Nov 25 14:07:45 crc kubenswrapper[4702]: E1125 14:07:45.776432 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="617d6aec-6354-4119-b0bd-48f71ba9b336" containerName="mariadb-account-delete" Nov 25 14:07:45 crc kubenswrapper[4702]: I1125 14:07:45.776501 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="617d6aec-6354-4119-b0bd-48f71ba9b336" containerName="mariadb-account-delete" Nov 25 14:07:45 crc kubenswrapper[4702]: I1125 14:07:45.776683 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="74a328c5-c8cd-49e6-ad80-de2b34117428" containerName="glance-log" Nov 25 14:07:45 crc kubenswrapper[4702]: I1125 14:07:45.776754 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="74a328c5-c8cd-49e6-ad80-de2b34117428" containerName="glance-httpd" Nov 25 14:07:45 crc kubenswrapper[4702]: I1125 14:07:45.776841 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="617d6aec-6354-4119-b0bd-48f71ba9b336" containerName="mariadb-account-delete" Nov 25 14:07:45 crc kubenswrapper[4702]: I1125 14:07:45.777515 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-zbqwl" Nov 25 14:07:45 crc kubenswrapper[4702]: I1125 14:07:45.784800 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a261226b-a600-4a09-a99f-1d886ddf8e6f-operator-scripts\") pod \"glance-db-create-zbqwl\" (UID: \"a261226b-a600-4a09-a99f-1d886ddf8e6f\") " pod="glance-kuttl-tests/glance-db-create-zbqwl" Nov 25 14:07:45 crc kubenswrapper[4702]: I1125 14:07:45.784938 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wf6k\" (UniqueName: \"kubernetes.io/projected/a261226b-a600-4a09-a99f-1d886ddf8e6f-kube-api-access-6wf6k\") pod \"glance-db-create-zbqwl\" (UID: \"a261226b-a600-4a09-a99f-1d886ddf8e6f\") " pod="glance-kuttl-tests/glance-db-create-zbqwl" Nov 25 14:07:45 crc kubenswrapper[4702]: I1125 14:07:45.787638 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-9a19-account-create-update-g98xf"] Nov 25 14:07:45 crc kubenswrapper[4702]: I1125 14:07:45.788878 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-9a19-account-create-update-g98xf" Nov 25 14:07:45 crc kubenswrapper[4702]: I1125 14:07:45.790951 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 25 14:07:45 crc kubenswrapper[4702]: I1125 14:07:45.797638 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-zbqwl"] Nov 25 14:07:45 crc kubenswrapper[4702]: I1125 14:07:45.815606 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="617d6aec-6354-4119-b0bd-48f71ba9b336" path="/var/lib/kubelet/pods/617d6aec-6354-4119-b0bd-48f71ba9b336/volumes" Nov 25 14:07:45 crc kubenswrapper[4702]: I1125 14:07:45.816494 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65a7b32a-6d99-4e03-92e7-a2d4f9e44441" path="/var/lib/kubelet/pods/65a7b32a-6d99-4e03-92e7-a2d4f9e44441/volumes" Nov 25 14:07:45 crc kubenswrapper[4702]: I1125 14:07:45.817108 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73b0c1bd-c025-419a-8a97-03523b6844b2" path="/var/lib/kubelet/pods/73b0c1bd-c025-419a-8a97-03523b6844b2/volumes" Nov 25 14:07:45 crc kubenswrapper[4702]: I1125 14:07:45.818442 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74a328c5-c8cd-49e6-ad80-de2b34117428" path="/var/lib/kubelet/pods/74a328c5-c8cd-49e6-ad80-de2b34117428/volumes" Nov 25 14:07:45 crc kubenswrapper[4702]: I1125 14:07:45.819039 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-9a19-account-create-update-g98xf"] Nov 25 14:07:45 crc kubenswrapper[4702]: I1125 14:07:45.885648 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe67604e-401d-4050-8c57-ed41d56b72f0-operator-scripts\") pod \"glance-9a19-account-create-update-g98xf\" (UID: \"fe67604e-401d-4050-8c57-ed41d56b72f0\") " pod="glance-kuttl-tests/glance-9a19-account-create-update-g98xf" Nov 25 14:07:45 crc kubenswrapper[4702]: I1125 14:07:45.885791 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a261226b-a600-4a09-a99f-1d886ddf8e6f-operator-scripts\") pod \"glance-db-create-zbqwl\" (UID: \"a261226b-a600-4a09-a99f-1d886ddf8e6f\") " pod="glance-kuttl-tests/glance-db-create-zbqwl" Nov 25 14:07:45 crc kubenswrapper[4702]: I1125 14:07:45.885828 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wf6k\" (UniqueName: \"kubernetes.io/projected/a261226b-a600-4a09-a99f-1d886ddf8e6f-kube-api-access-6wf6k\") pod \"glance-db-create-zbqwl\" (UID: \"a261226b-a600-4a09-a99f-1d886ddf8e6f\") " pod="glance-kuttl-tests/glance-db-create-zbqwl" Nov 25 14:07:45 crc kubenswrapper[4702]: I1125 14:07:45.885855 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wg6qx\" (UniqueName: \"kubernetes.io/projected/fe67604e-401d-4050-8c57-ed41d56b72f0-kube-api-access-wg6qx\") pod \"glance-9a19-account-create-update-g98xf\" (UID: \"fe67604e-401d-4050-8c57-ed41d56b72f0\") " pod="glance-kuttl-tests/glance-9a19-account-create-update-g98xf" Nov 25 14:07:45 crc kubenswrapper[4702]: I1125 14:07:45.886898 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a261226b-a600-4a09-a99f-1d886ddf8e6f-operator-scripts\") pod \"glance-db-create-zbqwl\" (UID: \"a261226b-a600-4a09-a99f-1d886ddf8e6f\") " pod="glance-kuttl-tests/glance-db-create-zbqwl" Nov 25 14:07:45 crc kubenswrapper[4702]: I1125 14:07:45.912417 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wf6k\" (UniqueName: \"kubernetes.io/projected/a261226b-a600-4a09-a99f-1d886ddf8e6f-kube-api-access-6wf6k\") pod \"glance-db-create-zbqwl\" (UID: \"a261226b-a600-4a09-a99f-1d886ddf8e6f\") " pod="glance-kuttl-tests/glance-db-create-zbqwl" Nov 25 14:07:45 crc kubenswrapper[4702]: I1125 14:07:45.987797 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wg6qx\" (UniqueName: \"kubernetes.io/projected/fe67604e-401d-4050-8c57-ed41d56b72f0-kube-api-access-wg6qx\") pod \"glance-9a19-account-create-update-g98xf\" (UID: \"fe67604e-401d-4050-8c57-ed41d56b72f0\") " pod="glance-kuttl-tests/glance-9a19-account-create-update-g98xf" Nov 25 14:07:45 crc kubenswrapper[4702]: I1125 14:07:45.987959 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe67604e-401d-4050-8c57-ed41d56b72f0-operator-scripts\") pod \"glance-9a19-account-create-update-g98xf\" (UID: \"fe67604e-401d-4050-8c57-ed41d56b72f0\") " pod="glance-kuttl-tests/glance-9a19-account-create-update-g98xf" Nov 25 14:07:45 crc kubenswrapper[4702]: I1125 14:07:45.989396 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe67604e-401d-4050-8c57-ed41d56b72f0-operator-scripts\") pod \"glance-9a19-account-create-update-g98xf\" (UID: \"fe67604e-401d-4050-8c57-ed41d56b72f0\") " pod="glance-kuttl-tests/glance-9a19-account-create-update-g98xf" Nov 25 14:07:46 crc kubenswrapper[4702]: I1125 14:07:46.006823 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wg6qx\" (UniqueName: \"kubernetes.io/projected/fe67604e-401d-4050-8c57-ed41d56b72f0-kube-api-access-wg6qx\") pod \"glance-9a19-account-create-update-g98xf\" (UID: \"fe67604e-401d-4050-8c57-ed41d56b72f0\") " pod="glance-kuttl-tests/glance-9a19-account-create-update-g98xf" Nov 25 14:07:46 crc kubenswrapper[4702]: I1125 14:07:46.104132 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-zbqwl" Nov 25 14:07:46 crc kubenswrapper[4702]: I1125 14:07:46.112947 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-9a19-account-create-update-g98xf" Nov 25 14:07:49 crc kubenswrapper[4702]: I1125 14:07:49.154369 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-zbqwl"] Nov 25 14:07:49 crc kubenswrapper[4702]: I1125 14:07:49.246395 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-9a19-account-create-update-g98xf"] Nov 25 14:07:49 crc kubenswrapper[4702]: I1125 14:07:49.774340 4702 generic.go:334] "Generic (PLEG): container finished" podID="a261226b-a600-4a09-a99f-1d886ddf8e6f" containerID="b9e3730cdee2ae8bac9ed922e1d63339550bc037a43007735d07fcd58c75e95e" exitCode=0 Nov 25 14:07:49 crc kubenswrapper[4702]: I1125 14:07:49.774665 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-zbqwl" event={"ID":"a261226b-a600-4a09-a99f-1d886ddf8e6f","Type":"ContainerDied","Data":"b9e3730cdee2ae8bac9ed922e1d63339550bc037a43007735d07fcd58c75e95e"} Nov 25 14:07:49 crc kubenswrapper[4702]: I1125 14:07:49.775007 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-zbqwl" event={"ID":"a261226b-a600-4a09-a99f-1d886ddf8e6f","Type":"ContainerStarted","Data":"ca86c387bb96e8002217ae1cde0f0cf4125a63ab4ecc219e5ffff5e0db320bfd"} Nov 25 14:07:49 crc kubenswrapper[4702]: I1125 14:07:49.777987 4702 generic.go:334] "Generic (PLEG): container finished" podID="fe67604e-401d-4050-8c57-ed41d56b72f0" containerID="ce06422e8fcd751bfcda6a78556c07b9613857458992640ca90dce8f564cdb7d" exitCode=0 Nov 25 14:07:49 crc kubenswrapper[4702]: I1125 14:07:49.778022 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-9a19-account-create-update-g98xf" event={"ID":"fe67604e-401d-4050-8c57-ed41d56b72f0","Type":"ContainerDied","Data":"ce06422e8fcd751bfcda6a78556c07b9613857458992640ca90dce8f564cdb7d"} Nov 25 14:07:49 crc kubenswrapper[4702]: I1125 14:07:49.778044 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-9a19-account-create-update-g98xf" event={"ID":"fe67604e-401d-4050-8c57-ed41d56b72f0","Type":"ContainerStarted","Data":"3eb32c0ba6b56bcd50a6fee992c6c9bdd163fd5ef10736b7440ca10b8c010e2c"} Nov 25 14:07:51 crc kubenswrapper[4702]: I1125 14:07:51.143593 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-zbqwl" Nov 25 14:07:51 crc kubenswrapper[4702]: I1125 14:07:51.148956 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-9a19-account-create-update-g98xf" Nov 25 14:07:51 crc kubenswrapper[4702]: I1125 14:07:51.269712 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe67604e-401d-4050-8c57-ed41d56b72f0-operator-scripts\") pod \"fe67604e-401d-4050-8c57-ed41d56b72f0\" (UID: \"fe67604e-401d-4050-8c57-ed41d56b72f0\") " Nov 25 14:07:51 crc kubenswrapper[4702]: I1125 14:07:51.270195 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wg6qx\" (UniqueName: \"kubernetes.io/projected/fe67604e-401d-4050-8c57-ed41d56b72f0-kube-api-access-wg6qx\") pod \"fe67604e-401d-4050-8c57-ed41d56b72f0\" (UID: \"fe67604e-401d-4050-8c57-ed41d56b72f0\") " Nov 25 14:07:51 crc kubenswrapper[4702]: I1125 14:07:51.270298 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6wf6k\" (UniqueName: \"kubernetes.io/projected/a261226b-a600-4a09-a99f-1d886ddf8e6f-kube-api-access-6wf6k\") pod \"a261226b-a600-4a09-a99f-1d886ddf8e6f\" (UID: \"a261226b-a600-4a09-a99f-1d886ddf8e6f\") " Nov 25 14:07:51 crc kubenswrapper[4702]: I1125 14:07:51.270374 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a261226b-a600-4a09-a99f-1d886ddf8e6f-operator-scripts\") pod \"a261226b-a600-4a09-a99f-1d886ddf8e6f\" (UID: \"a261226b-a600-4a09-a99f-1d886ddf8e6f\") " Nov 25 14:07:51 crc kubenswrapper[4702]: I1125 14:07:51.270813 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe67604e-401d-4050-8c57-ed41d56b72f0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fe67604e-401d-4050-8c57-ed41d56b72f0" (UID: "fe67604e-401d-4050-8c57-ed41d56b72f0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 14:07:51 crc kubenswrapper[4702]: I1125 14:07:51.271065 4702 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe67604e-401d-4050-8c57-ed41d56b72f0-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:51 crc kubenswrapper[4702]: I1125 14:07:51.271683 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a261226b-a600-4a09-a99f-1d886ddf8e6f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a261226b-a600-4a09-a99f-1d886ddf8e6f" (UID: "a261226b-a600-4a09-a99f-1d886ddf8e6f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 14:07:51 crc kubenswrapper[4702]: I1125 14:07:51.277437 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe67604e-401d-4050-8c57-ed41d56b72f0-kube-api-access-wg6qx" (OuterVolumeSpecName: "kube-api-access-wg6qx") pod "fe67604e-401d-4050-8c57-ed41d56b72f0" (UID: "fe67604e-401d-4050-8c57-ed41d56b72f0"). InnerVolumeSpecName "kube-api-access-wg6qx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:07:51 crc kubenswrapper[4702]: I1125 14:07:51.277553 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a261226b-a600-4a09-a99f-1d886ddf8e6f-kube-api-access-6wf6k" (OuterVolumeSpecName: "kube-api-access-6wf6k") pod "a261226b-a600-4a09-a99f-1d886ddf8e6f" (UID: "a261226b-a600-4a09-a99f-1d886ddf8e6f"). InnerVolumeSpecName "kube-api-access-6wf6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:07:51 crc kubenswrapper[4702]: I1125 14:07:51.373318 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wg6qx\" (UniqueName: \"kubernetes.io/projected/fe67604e-401d-4050-8c57-ed41d56b72f0-kube-api-access-wg6qx\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:51 crc kubenswrapper[4702]: I1125 14:07:51.373360 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6wf6k\" (UniqueName: \"kubernetes.io/projected/a261226b-a600-4a09-a99f-1d886ddf8e6f-kube-api-access-6wf6k\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:51 crc kubenswrapper[4702]: I1125 14:07:51.373372 4702 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a261226b-a600-4a09-a99f-1d886ddf8e6f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 14:07:51 crc kubenswrapper[4702]: I1125 14:07:51.813385 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-9a19-account-create-update-g98xf" Nov 25 14:07:51 crc kubenswrapper[4702]: I1125 14:07:51.814926 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-zbqwl" Nov 25 14:07:51 crc kubenswrapper[4702]: I1125 14:07:51.816361 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-9a19-account-create-update-g98xf" event={"ID":"fe67604e-401d-4050-8c57-ed41d56b72f0","Type":"ContainerDied","Data":"3eb32c0ba6b56bcd50a6fee992c6c9bdd163fd5ef10736b7440ca10b8c010e2c"} Nov 25 14:07:51 crc kubenswrapper[4702]: I1125 14:07:51.816463 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3eb32c0ba6b56bcd50a6fee992c6c9bdd163fd5ef10736b7440ca10b8c010e2c" Nov 25 14:07:51 crc kubenswrapper[4702]: I1125 14:07:51.816517 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-zbqwl" event={"ID":"a261226b-a600-4a09-a99f-1d886ddf8e6f","Type":"ContainerDied","Data":"ca86c387bb96e8002217ae1cde0f0cf4125a63ab4ecc219e5ffff5e0db320bfd"} Nov 25 14:07:51 crc kubenswrapper[4702]: I1125 14:07:51.816570 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca86c387bb96e8002217ae1cde0f0cf4125a63ab4ecc219e5ffff5e0db320bfd" Nov 25 14:07:56 crc kubenswrapper[4702]: I1125 14:07:56.036044 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-zmh5f"] Nov 25 14:07:56 crc kubenswrapper[4702]: E1125 14:07:56.036674 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a261226b-a600-4a09-a99f-1d886ddf8e6f" containerName="mariadb-database-create" Nov 25 14:07:56 crc kubenswrapper[4702]: I1125 14:07:56.036689 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="a261226b-a600-4a09-a99f-1d886ddf8e6f" containerName="mariadb-database-create" Nov 25 14:07:56 crc kubenswrapper[4702]: E1125 14:07:56.036724 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe67604e-401d-4050-8c57-ed41d56b72f0" containerName="mariadb-account-create-update" Nov 25 14:07:56 crc kubenswrapper[4702]: I1125 14:07:56.036730 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe67604e-401d-4050-8c57-ed41d56b72f0" containerName="mariadb-account-create-update" Nov 25 14:07:56 crc kubenswrapper[4702]: I1125 14:07:56.036851 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="a261226b-a600-4a09-a99f-1d886ddf8e6f" containerName="mariadb-database-create" Nov 25 14:07:56 crc kubenswrapper[4702]: I1125 14:07:56.036866 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe67604e-401d-4050-8c57-ed41d56b72f0" containerName="mariadb-account-create-update" Nov 25 14:07:56 crc kubenswrapper[4702]: I1125 14:07:56.037370 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-zmh5f" Nov 25 14:07:56 crc kubenswrapper[4702]: I1125 14:07:56.045649 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 25 14:07:56 crc kubenswrapper[4702]: I1125 14:07:56.045215 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-zmh5f"] Nov 25 14:07:56 crc kubenswrapper[4702]: I1125 14:07:56.045990 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-2tcvf" Nov 25 14:07:56 crc kubenswrapper[4702]: I1125 14:07:56.153111 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cf33d671-2bc3-46aa-95e9-eceb37bfe864-db-sync-config-data\") pod \"glance-db-sync-zmh5f\" (UID: \"cf33d671-2bc3-46aa-95e9-eceb37bfe864\") " pod="glance-kuttl-tests/glance-db-sync-zmh5f" Nov 25 14:07:56 crc kubenswrapper[4702]: I1125 14:07:56.153172 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfnqh\" (UniqueName: \"kubernetes.io/projected/cf33d671-2bc3-46aa-95e9-eceb37bfe864-kube-api-access-pfnqh\") pod \"glance-db-sync-zmh5f\" (UID: \"cf33d671-2bc3-46aa-95e9-eceb37bfe864\") " pod="glance-kuttl-tests/glance-db-sync-zmh5f" Nov 25 14:07:56 crc kubenswrapper[4702]: I1125 14:07:56.153427 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf33d671-2bc3-46aa-95e9-eceb37bfe864-config-data\") pod \"glance-db-sync-zmh5f\" (UID: \"cf33d671-2bc3-46aa-95e9-eceb37bfe864\") " pod="glance-kuttl-tests/glance-db-sync-zmh5f" Nov 25 14:07:56 crc kubenswrapper[4702]: I1125 14:07:56.255510 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cf33d671-2bc3-46aa-95e9-eceb37bfe864-db-sync-config-data\") pod \"glance-db-sync-zmh5f\" (UID: \"cf33d671-2bc3-46aa-95e9-eceb37bfe864\") " pod="glance-kuttl-tests/glance-db-sync-zmh5f" Nov 25 14:07:56 crc kubenswrapper[4702]: I1125 14:07:56.255561 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfnqh\" (UniqueName: \"kubernetes.io/projected/cf33d671-2bc3-46aa-95e9-eceb37bfe864-kube-api-access-pfnqh\") pod \"glance-db-sync-zmh5f\" (UID: \"cf33d671-2bc3-46aa-95e9-eceb37bfe864\") " pod="glance-kuttl-tests/glance-db-sync-zmh5f" Nov 25 14:07:56 crc kubenswrapper[4702]: I1125 14:07:56.255656 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf33d671-2bc3-46aa-95e9-eceb37bfe864-config-data\") pod \"glance-db-sync-zmh5f\" (UID: \"cf33d671-2bc3-46aa-95e9-eceb37bfe864\") " pod="glance-kuttl-tests/glance-db-sync-zmh5f" Nov 25 14:07:56 crc kubenswrapper[4702]: I1125 14:07:56.274338 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfnqh\" (UniqueName: \"kubernetes.io/projected/cf33d671-2bc3-46aa-95e9-eceb37bfe864-kube-api-access-pfnqh\") pod \"glance-db-sync-zmh5f\" (UID: \"cf33d671-2bc3-46aa-95e9-eceb37bfe864\") " pod="glance-kuttl-tests/glance-db-sync-zmh5f" Nov 25 14:07:56 crc kubenswrapper[4702]: I1125 14:07:56.274390 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cf33d671-2bc3-46aa-95e9-eceb37bfe864-db-sync-config-data\") pod \"glance-db-sync-zmh5f\" (UID: \"cf33d671-2bc3-46aa-95e9-eceb37bfe864\") " pod="glance-kuttl-tests/glance-db-sync-zmh5f" Nov 25 14:07:56 crc kubenswrapper[4702]: I1125 14:07:56.275960 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf33d671-2bc3-46aa-95e9-eceb37bfe864-config-data\") pod \"glance-db-sync-zmh5f\" (UID: \"cf33d671-2bc3-46aa-95e9-eceb37bfe864\") " pod="glance-kuttl-tests/glance-db-sync-zmh5f" Nov 25 14:07:56 crc kubenswrapper[4702]: I1125 14:07:56.361853 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-zmh5f" Nov 25 14:07:56 crc kubenswrapper[4702]: I1125 14:07:56.796168 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-zmh5f"] Nov 25 14:07:56 crc kubenswrapper[4702]: I1125 14:07:56.858035 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-zmh5f" event={"ID":"cf33d671-2bc3-46aa-95e9-eceb37bfe864","Type":"ContainerStarted","Data":"1739d9005766532600647bda0ea6e9516e5d64b92d7d66daee92d1678341d915"} Nov 25 14:07:57 crc kubenswrapper[4702]: I1125 14:07:57.866701 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-zmh5f" event={"ID":"cf33d671-2bc3-46aa-95e9-eceb37bfe864","Type":"ContainerStarted","Data":"4c7e021df80ed355ad656829dbe1d50152175b26e8f73f74f792b7105673d42f"} Nov 25 14:07:57 crc kubenswrapper[4702]: I1125 14:07:57.887508 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-zmh5f" podStartSLOduration=1.887482446 podStartE2EDuration="1.887482446s" podCreationTimestamp="2025-11-25 14:07:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 14:07:57.884485881 +0000 UTC m=+1287.047136331" watchObservedRunningTime="2025-11-25 14:07:57.887482446 +0000 UTC m=+1287.050132896" Nov 25 14:08:00 crc kubenswrapper[4702]: I1125 14:08:00.894424 4702 generic.go:334] "Generic (PLEG): container finished" podID="cf33d671-2bc3-46aa-95e9-eceb37bfe864" containerID="4c7e021df80ed355ad656829dbe1d50152175b26e8f73f74f792b7105673d42f" exitCode=0 Nov 25 14:08:00 crc kubenswrapper[4702]: I1125 14:08:00.894594 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-zmh5f" event={"ID":"cf33d671-2bc3-46aa-95e9-eceb37bfe864","Type":"ContainerDied","Data":"4c7e021df80ed355ad656829dbe1d50152175b26e8f73f74f792b7105673d42f"} Nov 25 14:08:02 crc kubenswrapper[4702]: I1125 14:08:02.266051 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-zmh5f" Nov 25 14:08:02 crc kubenswrapper[4702]: I1125 14:08:02.353750 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf33d671-2bc3-46aa-95e9-eceb37bfe864-config-data\") pod \"cf33d671-2bc3-46aa-95e9-eceb37bfe864\" (UID: \"cf33d671-2bc3-46aa-95e9-eceb37bfe864\") " Nov 25 14:08:02 crc kubenswrapper[4702]: I1125 14:08:02.353964 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cf33d671-2bc3-46aa-95e9-eceb37bfe864-db-sync-config-data\") pod \"cf33d671-2bc3-46aa-95e9-eceb37bfe864\" (UID: \"cf33d671-2bc3-46aa-95e9-eceb37bfe864\") " Nov 25 14:08:02 crc kubenswrapper[4702]: I1125 14:08:02.354079 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfnqh\" (UniqueName: \"kubernetes.io/projected/cf33d671-2bc3-46aa-95e9-eceb37bfe864-kube-api-access-pfnqh\") pod \"cf33d671-2bc3-46aa-95e9-eceb37bfe864\" (UID: \"cf33d671-2bc3-46aa-95e9-eceb37bfe864\") " Nov 25 14:08:02 crc kubenswrapper[4702]: I1125 14:08:02.363939 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf33d671-2bc3-46aa-95e9-eceb37bfe864-kube-api-access-pfnqh" (OuterVolumeSpecName: "kube-api-access-pfnqh") pod "cf33d671-2bc3-46aa-95e9-eceb37bfe864" (UID: "cf33d671-2bc3-46aa-95e9-eceb37bfe864"). InnerVolumeSpecName "kube-api-access-pfnqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:08:02 crc kubenswrapper[4702]: I1125 14:08:02.368367 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf33d671-2bc3-46aa-95e9-eceb37bfe864-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "cf33d671-2bc3-46aa-95e9-eceb37bfe864" (UID: "cf33d671-2bc3-46aa-95e9-eceb37bfe864"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:08:02 crc kubenswrapper[4702]: I1125 14:08:02.428718 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf33d671-2bc3-46aa-95e9-eceb37bfe864-config-data" (OuterVolumeSpecName: "config-data") pod "cf33d671-2bc3-46aa-95e9-eceb37bfe864" (UID: "cf33d671-2bc3-46aa-95e9-eceb37bfe864"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:08:02 crc kubenswrapper[4702]: I1125 14:08:02.456173 4702 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf33d671-2bc3-46aa-95e9-eceb37bfe864-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 14:08:02 crc kubenswrapper[4702]: I1125 14:08:02.456214 4702 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cf33d671-2bc3-46aa-95e9-eceb37bfe864-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 14:08:02 crc kubenswrapper[4702]: I1125 14:08:02.456227 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfnqh\" (UniqueName: \"kubernetes.io/projected/cf33d671-2bc3-46aa-95e9-eceb37bfe864-kube-api-access-pfnqh\") on node \"crc\" DevicePath \"\"" Nov 25 14:08:02 crc kubenswrapper[4702]: I1125 14:08:02.911809 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-zmh5f" event={"ID":"cf33d671-2bc3-46aa-95e9-eceb37bfe864","Type":"ContainerDied","Data":"1739d9005766532600647bda0ea6e9516e5d64b92d7d66daee92d1678341d915"} Nov 25 14:08:02 crc kubenswrapper[4702]: I1125 14:08:02.911862 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1739d9005766532600647bda0ea6e9516e5d64b92d7d66daee92d1678341d915" Nov 25 14:08:02 crc kubenswrapper[4702]: I1125 14:08:02.911890 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-zmh5f" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.332924 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 14:08:04 crc kubenswrapper[4702]: E1125 14:08:04.333619 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf33d671-2bc3-46aa-95e9-eceb37bfe864" containerName="glance-db-sync" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.333634 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf33d671-2bc3-46aa-95e9-eceb37bfe864" containerName="glance-db-sync" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.333781 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf33d671-2bc3-46aa-95e9-eceb37bfe864" containerName="glance-db-sync" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.334801 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.339687 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.340204 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-external-config-data" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.355836 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-2tcvf" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.407296 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.489983 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.491571 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.497046 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.513532 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.514523 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-dev\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.514565 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ba715ae-64e0-4810-b14d-f0c73b02f65d-scripts\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.514617 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ba715ae-64e0-4810-b14d-f0c73b02f65d-config-data\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.514659 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.514691 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.514711 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.514732 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2ba715ae-64e0-4810-b14d-f0c73b02f65d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.514753 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ba715ae-64e0-4810-b14d-f0c73b02f65d-logs\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.514776 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.514797 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-run\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.514824 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.514882 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbxxm\" (UniqueName: \"kubernetes.io/projected/2ba715ae-64e0-4810-b14d-f0c73b02f65d-kube-api-access-rbxxm\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.514908 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.514928 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-sys\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.616541 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.616635 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-dev\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.616665 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa1fb684-8de9-4737-888e-88f8e6021f6c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.616700 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ba715ae-64e0-4810-b14d-f0c73b02f65d-scripts\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.616728 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.616747 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.616775 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-dev\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.616808 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-sys\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.616831 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ba715ae-64e0-4810-b14d-f0c73b02f65d-config-data\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.616855 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.616882 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.616911 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.616932 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2ba715ae-64e0-4810-b14d-f0c73b02f65d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.616955 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aa1fb684-8de9-4737-888e-88f8e6021f6c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.616980 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrlr5\" (UniqueName: \"kubernetes.io/projected/aa1fb684-8de9-4737-888e-88f8e6021f6c-kube-api-access-lrlr5\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.617007 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.617033 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.617056 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ba715ae-64e0-4810-b14d-f0c73b02f65d-logs\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.617080 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.617105 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-run\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.617136 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.617159 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa1fb684-8de9-4737-888e-88f8e6021f6c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.617184 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa1fb684-8de9-4737-888e-88f8e6021f6c-logs\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.617262 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbxxm\" (UniqueName: \"kubernetes.io/projected/2ba715ae-64e0-4810-b14d-f0c73b02f65d-kube-api-access-rbxxm\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.617298 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.617341 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-sys\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.617358 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.617385 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-run\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.617486 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-dev\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.618361 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.618413 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-sys\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.618620 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.618700 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-run\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.618773 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.618814 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.619090 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.619167 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.619314 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2ba715ae-64e0-4810-b14d-f0c73b02f65d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.619337 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ba715ae-64e0-4810-b14d-f0c73b02f65d-logs\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.627413 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ba715ae-64e0-4810-b14d-f0c73b02f65d-scripts\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.642303 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ba715ae-64e0-4810-b14d-f0c73b02f65d-config-data\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.643985 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbxxm\" (UniqueName: \"kubernetes.io/projected/2ba715ae-64e0-4810-b14d-f0c73b02f65d-kube-api-access-rbxxm\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.644039 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.648469 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.718780 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-sys\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.718855 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.718886 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aa1fb684-8de9-4737-888e-88f8e6021f6c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.718955 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrlr5\" (UniqueName: \"kubernetes.io/projected/aa1fb684-8de9-4737-888e-88f8e6021f6c-kube-api-access-lrlr5\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.718982 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.719012 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa1fb684-8de9-4737-888e-88f8e6021f6c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.719031 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa1fb684-8de9-4737-888e-88f8e6021f6c-logs\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.719071 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.719098 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-run\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.719128 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.719162 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa1fb684-8de9-4737-888e-88f8e6021f6c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.719188 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.719203 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.719227 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-dev\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.719322 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-dev\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.719364 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-sys\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.719522 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.719995 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.720121 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.720065 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.720014 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-run\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.720282 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aa1fb684-8de9-4737-888e-88f8e6021f6c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.720381 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.720384 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.720724 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa1fb684-8de9-4737-888e-88f8e6021f6c-logs\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.731344 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa1fb684-8de9-4737-888e-88f8e6021f6c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.736255 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa1fb684-8de9-4737-888e-88f8e6021f6c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.762784 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrlr5\" (UniqueName: \"kubernetes.io/projected/aa1fb684-8de9-4737-888e-88f8e6021f6c-kube-api-access-lrlr5\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.772460 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.772628 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.811649 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:04 crc kubenswrapper[4702]: I1125 14:08:04.952897 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:05 crc kubenswrapper[4702]: I1125 14:08:05.366757 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 14:08:05 crc kubenswrapper[4702]: I1125 14:08:05.462577 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 14:08:05 crc kubenswrapper[4702]: W1125 14:08:05.467500 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ba715ae_64e0_4810_b14d_f0c73b02f65d.slice/crio-4c1fd19dc3bd9fc157489944a459533d90c6ee14297b20c83e64ee525e00ebde WatchSource:0}: Error finding container 4c1fd19dc3bd9fc157489944a459533d90c6ee14297b20c83e64ee525e00ebde: Status 404 returned error can't find the container with id 4c1fd19dc3bd9fc157489944a459533d90c6ee14297b20c83e64ee525e00ebde Nov 25 14:08:05 crc kubenswrapper[4702]: I1125 14:08:05.942602 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"aa1fb684-8de9-4737-888e-88f8e6021f6c","Type":"ContainerStarted","Data":"3bab896862c848b1c64c7eeda42230af914fb7bed2b6395b5f1c0a7cc9db6e88"} Nov 25 14:08:05 crc kubenswrapper[4702]: I1125 14:08:05.942655 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"aa1fb684-8de9-4737-888e-88f8e6021f6c","Type":"ContainerStarted","Data":"d82b664a53b833fd2cfaa12933c6b7add417ad8049c98762a563c61769c1f889"} Nov 25 14:08:05 crc kubenswrapper[4702]: I1125 14:08:05.944651 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"2ba715ae-64e0-4810-b14d-f0c73b02f65d","Type":"ContainerStarted","Data":"094e6d8e70f0a27cf92b867a94892cc0e5547cae1c180d5a2e21ad71123c29ec"} Nov 25 14:08:05 crc kubenswrapper[4702]: I1125 14:08:05.944710 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"2ba715ae-64e0-4810-b14d-f0c73b02f65d","Type":"ContainerStarted","Data":"4c1fd19dc3bd9fc157489944a459533d90c6ee14297b20c83e64ee525e00ebde"} Nov 25 14:08:06 crc kubenswrapper[4702]: I1125 14:08:06.041088 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 14:08:06 crc kubenswrapper[4702]: I1125 14:08:06.953031 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"aa1fb684-8de9-4737-888e-88f8e6021f6c","Type":"ContainerStarted","Data":"c43b9aad9da1304fbc50da5e714b326baf40eae661205ccf2e612c218ca6284a"} Nov 25 14:08:06 crc kubenswrapper[4702]: I1125 14:08:06.953724 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"aa1fb684-8de9-4737-888e-88f8e6021f6c","Type":"ContainerStarted","Data":"b1f7b7456ba279f672a2382d73feb14cdacae3fded2195f84e82c9ae8a697492"} Nov 25 14:08:06 crc kubenswrapper[4702]: I1125 14:08:06.953876 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="aa1fb684-8de9-4737-888e-88f8e6021f6c" containerName="glance-log" containerID="cri-o://3bab896862c848b1c64c7eeda42230af914fb7bed2b6395b5f1c0a7cc9db6e88" gracePeriod=30 Nov 25 14:08:06 crc kubenswrapper[4702]: I1125 14:08:06.954420 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="aa1fb684-8de9-4737-888e-88f8e6021f6c" containerName="glance-api" containerID="cri-o://c43b9aad9da1304fbc50da5e714b326baf40eae661205ccf2e612c218ca6284a" gracePeriod=30 Nov 25 14:08:06 crc kubenswrapper[4702]: I1125 14:08:06.954467 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="aa1fb684-8de9-4737-888e-88f8e6021f6c" containerName="glance-httpd" containerID="cri-o://b1f7b7456ba279f672a2382d73feb14cdacae3fded2195f84e82c9ae8a697492" gracePeriod=30 Nov 25 14:08:06 crc kubenswrapper[4702]: I1125 14:08:06.957408 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"2ba715ae-64e0-4810-b14d-f0c73b02f65d","Type":"ContainerStarted","Data":"4cf74f8aeeaa938a660d0ca5bfdef3ebfaebf66bb0ef971a12b7ef0b5dda5e11"} Nov 25 14:08:06 crc kubenswrapper[4702]: I1125 14:08:06.957443 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"2ba715ae-64e0-4810-b14d-f0c73b02f65d","Type":"ContainerStarted","Data":"b07b7cb2e145043a8a4f4070f63ce4ef540385d4bdc101570ce60ffaf6ac41b1"} Nov 25 14:08:06 crc kubenswrapper[4702]: I1125 14:08:06.995386 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=3.995361425 podStartE2EDuration="3.995361425s" podCreationTimestamp="2025-11-25 14:08:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 14:08:06.989343177 +0000 UTC m=+1296.151993637" watchObservedRunningTime="2025-11-25 14:08:06.995361425 +0000 UTC m=+1296.158011875" Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.020120 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-0" podStartSLOduration=3.020100136 podStartE2EDuration="3.020100136s" podCreationTimestamp="2025-11-25 14:08:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 14:08:07.015513383 +0000 UTC m=+1296.178163833" watchObservedRunningTime="2025-11-25 14:08:07.020100136 +0000 UTC m=+1296.182750586" Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.413253 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.477346 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrlr5\" (UniqueName: \"kubernetes.io/projected/aa1fb684-8de9-4737-888e-88f8e6021f6c-kube-api-access-lrlr5\") pod \"aa1fb684-8de9-4737-888e-88f8e6021f6c\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.477736 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-dev\") pod \"aa1fb684-8de9-4737-888e-88f8e6021f6c\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.477778 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-var-locks-brick\") pod \"aa1fb684-8de9-4737-888e-88f8e6021f6c\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.477820 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa1fb684-8de9-4737-888e-88f8e6021f6c-logs\") pod \"aa1fb684-8de9-4737-888e-88f8e6021f6c\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.477858 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa1fb684-8de9-4737-888e-88f8e6021f6c-config-data\") pod \"aa1fb684-8de9-4737-888e-88f8e6021f6c\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.477900 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-etc-iscsi\") pod \"aa1fb684-8de9-4737-888e-88f8e6021f6c\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.477888 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-dev" (OuterVolumeSpecName: "dev") pod "aa1fb684-8de9-4737-888e-88f8e6021f6c" (UID: "aa1fb684-8de9-4737-888e-88f8e6021f6c"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.477901 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "aa1fb684-8de9-4737-888e-88f8e6021f6c" (UID: "aa1fb684-8de9-4737-888e-88f8e6021f6c"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.477975 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-etc-nvme\") pod \"aa1fb684-8de9-4737-888e-88f8e6021f6c\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.477998 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-sys\") pod \"aa1fb684-8de9-4737-888e-88f8e6021f6c\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.478056 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "aa1fb684-8de9-4737-888e-88f8e6021f6c" (UID: "aa1fb684-8de9-4737-888e-88f8e6021f6c"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.478096 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-lib-modules\") pod \"aa1fb684-8de9-4737-888e-88f8e6021f6c\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.478142 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "aa1fb684-8de9-4737-888e-88f8e6021f6c" (UID: "aa1fb684-8de9-4737-888e-88f8e6021f6c"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.478201 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "aa1fb684-8de9-4737-888e-88f8e6021f6c" (UID: "aa1fb684-8de9-4737-888e-88f8e6021f6c"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.478163 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-sys" (OuterVolumeSpecName: "sys") pod "aa1fb684-8de9-4737-888e-88f8e6021f6c" (UID: "aa1fb684-8de9-4737-888e-88f8e6021f6c"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.478176 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa1fb684-8de9-4737-888e-88f8e6021f6c-scripts\") pod \"aa1fb684-8de9-4737-888e-88f8e6021f6c\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.478363 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"aa1fb684-8de9-4737-888e-88f8e6021f6c\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.478426 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-run\") pod \"aa1fb684-8de9-4737-888e-88f8e6021f6c\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.478491 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aa1fb684-8de9-4737-888e-88f8e6021f6c-httpd-run\") pod \"aa1fb684-8de9-4737-888e-88f8e6021f6c\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.478520 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"aa1fb684-8de9-4737-888e-88f8e6021f6c\" (UID: \"aa1fb684-8de9-4737-888e-88f8e6021f6c\") " Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.478638 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-run" (OuterVolumeSpecName: "run") pod "aa1fb684-8de9-4737-888e-88f8e6021f6c" (UID: "aa1fb684-8de9-4737-888e-88f8e6021f6c"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.479184 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa1fb684-8de9-4737-888e-88f8e6021f6c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "aa1fb684-8de9-4737-888e-88f8e6021f6c" (UID: "aa1fb684-8de9-4737-888e-88f8e6021f6c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.479570 4702 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-run\") on node \"crc\" DevicePath \"\"" Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.479594 4702 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aa1fb684-8de9-4737-888e-88f8e6021f6c-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.479609 4702 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-dev\") on node \"crc\" DevicePath \"\"" Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.479653 4702 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.479671 4702 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.479683 4702 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.479695 4702 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-sys\") on node \"crc\" DevicePath \"\"" Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.479707 4702 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/aa1fb684-8de9-4737-888e-88f8e6021f6c-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.480008 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa1fb684-8de9-4737-888e-88f8e6021f6c-logs" (OuterVolumeSpecName: "logs") pod "aa1fb684-8de9-4737-888e-88f8e6021f6c" (UID: "aa1fb684-8de9-4737-888e-88f8e6021f6c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.484488 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa1fb684-8de9-4737-888e-88f8e6021f6c-kube-api-access-lrlr5" (OuterVolumeSpecName: "kube-api-access-lrlr5") pod "aa1fb684-8de9-4737-888e-88f8e6021f6c" (UID: "aa1fb684-8de9-4737-888e-88f8e6021f6c"). InnerVolumeSpecName "kube-api-access-lrlr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.484771 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance-cache") pod "aa1fb684-8de9-4737-888e-88f8e6021f6c" (UID: "aa1fb684-8de9-4737-888e-88f8e6021f6c"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.486690 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "aa1fb684-8de9-4737-888e-88f8e6021f6c" (UID: "aa1fb684-8de9-4737-888e-88f8e6021f6c"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.489771 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa1fb684-8de9-4737-888e-88f8e6021f6c-scripts" (OuterVolumeSpecName: "scripts") pod "aa1fb684-8de9-4737-888e-88f8e6021f6c" (UID: "aa1fb684-8de9-4737-888e-88f8e6021f6c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.564847 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa1fb684-8de9-4737-888e-88f8e6021f6c-config-data" (OuterVolumeSpecName: "config-data") pod "aa1fb684-8de9-4737-888e-88f8e6021f6c" (UID: "aa1fb684-8de9-4737-888e-88f8e6021f6c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.581020 4702 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa1fb684-8de9-4737-888e-88f8e6021f6c-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.581107 4702 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.581157 4702 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.581172 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrlr5\" (UniqueName: \"kubernetes.io/projected/aa1fb684-8de9-4737-888e-88f8e6021f6c-kube-api-access-lrlr5\") on node \"crc\" DevicePath \"\"" Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.581186 4702 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa1fb684-8de9-4737-888e-88f8e6021f6c-logs\") on node \"crc\" DevicePath \"\"" Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.581202 4702 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa1fb684-8de9-4737-888e-88f8e6021f6c-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.595520 4702 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.598539 4702 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 25 14:08:07 crc kubenswrapper[4702]: W1125 14:08:07.608439 4702 watcher.go:93] Error while processing event ("/sys/fs/cgroup/user.slice/user-0.slice/session-c11.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/user.slice/user-0.slice/session-c11.scope: no such file or directory Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.683191 4702 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.683291 4702 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.968156 4702 generic.go:334] "Generic (PLEG): container finished" podID="aa1fb684-8de9-4737-888e-88f8e6021f6c" containerID="c43b9aad9da1304fbc50da5e714b326baf40eae661205ccf2e612c218ca6284a" exitCode=143 Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.968612 4702 generic.go:334] "Generic (PLEG): container finished" podID="aa1fb684-8de9-4737-888e-88f8e6021f6c" containerID="b1f7b7456ba279f672a2382d73feb14cdacae3fded2195f84e82c9ae8a697492" exitCode=143 Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.968622 4702 generic.go:334] "Generic (PLEG): container finished" podID="aa1fb684-8de9-4737-888e-88f8e6021f6c" containerID="3bab896862c848b1c64c7eeda42230af914fb7bed2b6395b5f1c0a7cc9db6e88" exitCode=143 Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.968254 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"aa1fb684-8de9-4737-888e-88f8e6021f6c","Type":"ContainerDied","Data":"c43b9aad9da1304fbc50da5e714b326baf40eae661205ccf2e612c218ca6284a"} Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.968694 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"aa1fb684-8de9-4737-888e-88f8e6021f6c","Type":"ContainerDied","Data":"b1f7b7456ba279f672a2382d73feb14cdacae3fded2195f84e82c9ae8a697492"} Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.968707 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"aa1fb684-8de9-4737-888e-88f8e6021f6c","Type":"ContainerDied","Data":"3bab896862c848b1c64c7eeda42230af914fb7bed2b6395b5f1c0a7cc9db6e88"} Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.968713 4702 scope.go:117] "RemoveContainer" containerID="c43b9aad9da1304fbc50da5e714b326baf40eae661205ccf2e612c218ca6284a" Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.968720 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"aa1fb684-8de9-4737-888e-88f8e6021f6c","Type":"ContainerDied","Data":"d82b664a53b833fd2cfaa12933c6b7add417ad8049c98762a563c61769c1f889"} Nov 25 14:08:07 crc kubenswrapper[4702]: I1125 14:08:07.968279 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.010663 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.011591 4702 scope.go:117] "RemoveContainer" containerID="b1f7b7456ba279f672a2382d73feb14cdacae3fded2195f84e82c9ae8a697492" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.017822 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.023894 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 14:08:08 crc kubenswrapper[4702]: E1125 14:08:08.024360 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa1fb684-8de9-4737-888e-88f8e6021f6c" containerName="glance-api" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.024385 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa1fb684-8de9-4737-888e-88f8e6021f6c" containerName="glance-api" Nov 25 14:08:08 crc kubenswrapper[4702]: E1125 14:08:08.024399 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa1fb684-8de9-4737-888e-88f8e6021f6c" containerName="glance-log" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.024409 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa1fb684-8de9-4737-888e-88f8e6021f6c" containerName="glance-log" Nov 25 14:08:08 crc kubenswrapper[4702]: E1125 14:08:08.024442 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa1fb684-8de9-4737-888e-88f8e6021f6c" containerName="glance-httpd" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.024449 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa1fb684-8de9-4737-888e-88f8e6021f6c" containerName="glance-httpd" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.024649 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa1fb684-8de9-4737-888e-88f8e6021f6c" containerName="glance-api" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.024665 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa1fb684-8de9-4737-888e-88f8e6021f6c" containerName="glance-httpd" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.024672 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa1fb684-8de9-4737-888e-88f8e6021f6c" containerName="glance-log" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.025701 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.028560 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.036564 4702 scope.go:117] "RemoveContainer" containerID="3bab896862c848b1c64c7eeda42230af914fb7bed2b6395b5f1c0a7cc9db6e88" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.044133 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.100564 4702 scope.go:117] "RemoveContainer" containerID="c43b9aad9da1304fbc50da5e714b326baf40eae661205ccf2e612c218ca6284a" Nov 25 14:08:08 crc kubenswrapper[4702]: E1125 14:08:08.103146 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c43b9aad9da1304fbc50da5e714b326baf40eae661205ccf2e612c218ca6284a\": container with ID starting with c43b9aad9da1304fbc50da5e714b326baf40eae661205ccf2e612c218ca6284a not found: ID does not exist" containerID="c43b9aad9da1304fbc50da5e714b326baf40eae661205ccf2e612c218ca6284a" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.103184 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c43b9aad9da1304fbc50da5e714b326baf40eae661205ccf2e612c218ca6284a"} err="failed to get container status \"c43b9aad9da1304fbc50da5e714b326baf40eae661205ccf2e612c218ca6284a\": rpc error: code = NotFound desc = could not find container \"c43b9aad9da1304fbc50da5e714b326baf40eae661205ccf2e612c218ca6284a\": container with ID starting with c43b9aad9da1304fbc50da5e714b326baf40eae661205ccf2e612c218ca6284a not found: ID does not exist" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.103210 4702 scope.go:117] "RemoveContainer" containerID="b1f7b7456ba279f672a2382d73feb14cdacae3fded2195f84e82c9ae8a697492" Nov 25 14:08:08 crc kubenswrapper[4702]: E1125 14:08:08.103573 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1f7b7456ba279f672a2382d73feb14cdacae3fded2195f84e82c9ae8a697492\": container with ID starting with b1f7b7456ba279f672a2382d73feb14cdacae3fded2195f84e82c9ae8a697492 not found: ID does not exist" containerID="b1f7b7456ba279f672a2382d73feb14cdacae3fded2195f84e82c9ae8a697492" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.103620 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1f7b7456ba279f672a2382d73feb14cdacae3fded2195f84e82c9ae8a697492"} err="failed to get container status \"b1f7b7456ba279f672a2382d73feb14cdacae3fded2195f84e82c9ae8a697492\": rpc error: code = NotFound desc = could not find container \"b1f7b7456ba279f672a2382d73feb14cdacae3fded2195f84e82c9ae8a697492\": container with ID starting with b1f7b7456ba279f672a2382d73feb14cdacae3fded2195f84e82c9ae8a697492 not found: ID does not exist" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.103651 4702 scope.go:117] "RemoveContainer" containerID="3bab896862c848b1c64c7eeda42230af914fb7bed2b6395b5f1c0a7cc9db6e88" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.104797 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-sys\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.104860 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-run\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.104881 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.104958 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.105033 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.105797 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.105824 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.105903 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-logs\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.105926 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcd6s\" (UniqueName: \"kubernetes.io/projected/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-kube-api-access-vcd6s\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.105954 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-dev\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.106027 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-scripts\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.106086 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-config-data\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.106104 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.106312 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: E1125 14:08:08.108268 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bab896862c848b1c64c7eeda42230af914fb7bed2b6395b5f1c0a7cc9db6e88\": container with ID starting with 3bab896862c848b1c64c7eeda42230af914fb7bed2b6395b5f1c0a7cc9db6e88 not found: ID does not exist" containerID="3bab896862c848b1c64c7eeda42230af914fb7bed2b6395b5f1c0a7cc9db6e88" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.108329 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bab896862c848b1c64c7eeda42230af914fb7bed2b6395b5f1c0a7cc9db6e88"} err="failed to get container status \"3bab896862c848b1c64c7eeda42230af914fb7bed2b6395b5f1c0a7cc9db6e88\": rpc error: code = NotFound desc = could not find container \"3bab896862c848b1c64c7eeda42230af914fb7bed2b6395b5f1c0a7cc9db6e88\": container with ID starting with 3bab896862c848b1c64c7eeda42230af914fb7bed2b6395b5f1c0a7cc9db6e88 not found: ID does not exist" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.108362 4702 scope.go:117] "RemoveContainer" containerID="c43b9aad9da1304fbc50da5e714b326baf40eae661205ccf2e612c218ca6284a" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.110440 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c43b9aad9da1304fbc50da5e714b326baf40eae661205ccf2e612c218ca6284a"} err="failed to get container status \"c43b9aad9da1304fbc50da5e714b326baf40eae661205ccf2e612c218ca6284a\": rpc error: code = NotFound desc = could not find container \"c43b9aad9da1304fbc50da5e714b326baf40eae661205ccf2e612c218ca6284a\": container with ID starting with c43b9aad9da1304fbc50da5e714b326baf40eae661205ccf2e612c218ca6284a not found: ID does not exist" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.110473 4702 scope.go:117] "RemoveContainer" containerID="b1f7b7456ba279f672a2382d73feb14cdacae3fded2195f84e82c9ae8a697492" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.110889 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1f7b7456ba279f672a2382d73feb14cdacae3fded2195f84e82c9ae8a697492"} err="failed to get container status \"b1f7b7456ba279f672a2382d73feb14cdacae3fded2195f84e82c9ae8a697492\": rpc error: code = NotFound desc = could not find container \"b1f7b7456ba279f672a2382d73feb14cdacae3fded2195f84e82c9ae8a697492\": container with ID starting with b1f7b7456ba279f672a2382d73feb14cdacae3fded2195f84e82c9ae8a697492 not found: ID does not exist" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.110909 4702 scope.go:117] "RemoveContainer" containerID="3bab896862c848b1c64c7eeda42230af914fb7bed2b6395b5f1c0a7cc9db6e88" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.111343 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bab896862c848b1c64c7eeda42230af914fb7bed2b6395b5f1c0a7cc9db6e88"} err="failed to get container status \"3bab896862c848b1c64c7eeda42230af914fb7bed2b6395b5f1c0a7cc9db6e88\": rpc error: code = NotFound desc = could not find container \"3bab896862c848b1c64c7eeda42230af914fb7bed2b6395b5f1c0a7cc9db6e88\": container with ID starting with 3bab896862c848b1c64c7eeda42230af914fb7bed2b6395b5f1c0a7cc9db6e88 not found: ID does not exist" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.111365 4702 scope.go:117] "RemoveContainer" containerID="c43b9aad9da1304fbc50da5e714b326baf40eae661205ccf2e612c218ca6284a" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.111578 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c43b9aad9da1304fbc50da5e714b326baf40eae661205ccf2e612c218ca6284a"} err="failed to get container status \"c43b9aad9da1304fbc50da5e714b326baf40eae661205ccf2e612c218ca6284a\": rpc error: code = NotFound desc = could not find container \"c43b9aad9da1304fbc50da5e714b326baf40eae661205ccf2e612c218ca6284a\": container with ID starting with c43b9aad9da1304fbc50da5e714b326baf40eae661205ccf2e612c218ca6284a not found: ID does not exist" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.111599 4702 scope.go:117] "RemoveContainer" containerID="b1f7b7456ba279f672a2382d73feb14cdacae3fded2195f84e82c9ae8a697492" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.111857 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1f7b7456ba279f672a2382d73feb14cdacae3fded2195f84e82c9ae8a697492"} err="failed to get container status \"b1f7b7456ba279f672a2382d73feb14cdacae3fded2195f84e82c9ae8a697492\": rpc error: code = NotFound desc = could not find container \"b1f7b7456ba279f672a2382d73feb14cdacae3fded2195f84e82c9ae8a697492\": container with ID starting with b1f7b7456ba279f672a2382d73feb14cdacae3fded2195f84e82c9ae8a697492 not found: ID does not exist" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.111877 4702 scope.go:117] "RemoveContainer" containerID="3bab896862c848b1c64c7eeda42230af914fb7bed2b6395b5f1c0a7cc9db6e88" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.112074 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bab896862c848b1c64c7eeda42230af914fb7bed2b6395b5f1c0a7cc9db6e88"} err="failed to get container status \"3bab896862c848b1c64c7eeda42230af914fb7bed2b6395b5f1c0a7cc9db6e88\": rpc error: code = NotFound desc = could not find container \"3bab896862c848b1c64c7eeda42230af914fb7bed2b6395b5f1c0a7cc9db6e88\": container with ID starting with 3bab896862c848b1c64c7eeda42230af914fb7bed2b6395b5f1c0a7cc9db6e88 not found: ID does not exist" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.207070 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-config-data\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.207108 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.207128 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-scripts\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.207154 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.207179 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-sys\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.207197 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-run\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.207216 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.207317 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.207362 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.207387 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.207407 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.207426 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-logs\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.207443 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcd6s\" (UniqueName: \"kubernetes.io/projected/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-kube-api-access-vcd6s\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.207465 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-dev\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.207541 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-dev\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.208557 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.209439 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.209458 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-sys\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.209504 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-run\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.209517 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.209557 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.209465 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.209552 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.209640 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.209786 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-logs\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.214010 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-config-data\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.227353 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcd6s\" (UniqueName: \"kubernetes.io/projected/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-kube-api-access-vcd6s\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.229979 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-scripts\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.240703 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.243758 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.393586 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.819788 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 14:08:08 crc kubenswrapper[4702]: I1125 14:08:08.978367 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"50e365be-ffbe-4bf9-8eec-19cca5ca20cb","Type":"ContainerStarted","Data":"fac158d959802d93193f8df09824e8bb237100e63101313bd6087057449c101b"} Nov 25 14:08:09 crc kubenswrapper[4702]: I1125 14:08:09.809652 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa1fb684-8de9-4737-888e-88f8e6021f6c" path="/var/lib/kubelet/pods/aa1fb684-8de9-4737-888e-88f8e6021f6c/volumes" Nov 25 14:08:09 crc kubenswrapper[4702]: I1125 14:08:09.989758 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"50e365be-ffbe-4bf9-8eec-19cca5ca20cb","Type":"ContainerStarted","Data":"5df1dfc91583f02a0504817678e09b0d330f585b91b0df332270cca5ebc47573"} Nov 25 14:08:09 crc kubenswrapper[4702]: I1125 14:08:09.989801 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"50e365be-ffbe-4bf9-8eec-19cca5ca20cb","Type":"ContainerStarted","Data":"f6440decba611fb6665099ea729c5d8b60b42b55644f2be04030b74123ceead2"} Nov 25 14:08:09 crc kubenswrapper[4702]: I1125 14:08:09.989812 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"50e365be-ffbe-4bf9-8eec-19cca5ca20cb","Type":"ContainerStarted","Data":"cd9ebd18fca903bcc5c3910acd17c494abdac265821f0092b48049e2d699a783"} Nov 25 14:08:10 crc kubenswrapper[4702]: I1125 14:08:10.018904 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=3.018885152 podStartE2EDuration="3.018885152s" podCreationTimestamp="2025-11-25 14:08:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 14:08:10.012545545 +0000 UTC m=+1299.175195995" watchObservedRunningTime="2025-11-25 14:08:10.018885152 +0000 UTC m=+1299.181535592" Nov 25 14:08:14 crc kubenswrapper[4702]: I1125 14:08:14.953011 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:14 crc kubenswrapper[4702]: I1125 14:08:14.953777 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:14 crc kubenswrapper[4702]: I1125 14:08:14.953793 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:14 crc kubenswrapper[4702]: I1125 14:08:14.979459 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:14 crc kubenswrapper[4702]: I1125 14:08:14.992005 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:14 crc kubenswrapper[4702]: I1125 14:08:14.999396 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:15 crc kubenswrapper[4702]: I1125 14:08:15.035068 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:15 crc kubenswrapper[4702]: I1125 14:08:15.035117 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:15 crc kubenswrapper[4702]: I1125 14:08:15.035126 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:15 crc kubenswrapper[4702]: I1125 14:08:15.064215 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:15 crc kubenswrapper[4702]: I1125 14:08:15.073002 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:15 crc kubenswrapper[4702]: I1125 14:08:15.075978 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:08:18 crc kubenswrapper[4702]: I1125 14:08:18.394646 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:18 crc kubenswrapper[4702]: I1125 14:08:18.394715 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:18 crc kubenswrapper[4702]: I1125 14:08:18.394728 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:18 crc kubenswrapper[4702]: I1125 14:08:18.423345 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:18 crc kubenswrapper[4702]: I1125 14:08:18.424920 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:18 crc kubenswrapper[4702]: I1125 14:08:18.446069 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:19 crc kubenswrapper[4702]: I1125 14:08:19.068678 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:19 crc kubenswrapper[4702]: I1125 14:08:19.068724 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:19 crc kubenswrapper[4702]: I1125 14:08:19.068739 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:19 crc kubenswrapper[4702]: I1125 14:08:19.082648 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:19 crc kubenswrapper[4702]: I1125 14:08:19.082891 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:19 crc kubenswrapper[4702]: I1125 14:08:19.096716 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:08:20 crc kubenswrapper[4702]: I1125 14:08:20.970455 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 25 14:08:20 crc kubenswrapper[4702]: I1125 14:08:20.989538 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 14:08:20 crc kubenswrapper[4702]: I1125 14:08:20.991996 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:20 crc kubenswrapper[4702]: I1125 14:08:20.993202 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.007925 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.015633 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.122745 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qp2fk\" (UniqueName: \"kubernetes.io/projected/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-kube-api-access-qp2fk\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.122798 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-run\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.122838 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.122856 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-sys\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.122878 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.123038 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.123087 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-dev\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.123156 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.123186 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-run\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.123267 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-logs\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.123319 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-scripts\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.123348 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.123407 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.123444 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.123469 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.123511 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-logs\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.123538 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.123566 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-sys\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.123585 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.123614 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-scripts\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.124718 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-config-data\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.124759 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.124786 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.124803 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.124839 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-config-data\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.124861 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2jwr\" (UniqueName: \"kubernetes.io/projected/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-kube-api-access-j2jwr\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.124897 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.124924 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-dev\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.226411 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2jwr\" (UniqueName: \"kubernetes.io/projected/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-kube-api-access-j2jwr\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.226477 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.226498 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-dev\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.226543 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qp2fk\" (UniqueName: \"kubernetes.io/projected/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-kube-api-access-qp2fk\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.226562 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-run\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.226595 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.226610 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-sys\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.226643 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.226665 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.226683 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-dev\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.226707 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.226728 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-run\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.226721 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-run\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.226823 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.226862 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.226897 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-sys\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.226752 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-logs\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.226938 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-dev\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.227001 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-scripts\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.227036 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.227067 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.227109 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.227138 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.227162 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-logs\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.227186 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.227198 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.227265 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.227293 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-sys\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.227315 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-run\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.227322 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.227326 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-logs\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.227348 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.227367 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-scripts\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.227389 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.227395 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-config-data\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.227439 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.227449 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.227483 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.227486 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.227530 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.227505 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.227642 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-config-data\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.227742 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.227855 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-logs\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.227976 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-sys\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.228003 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.228316 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.228321 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") device mount path \"/mnt/openstack/pv02\"" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.228609 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.228789 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-dev\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.236282 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-scripts\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.236919 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-config-data\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.238289 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-scripts\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.241486 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-config-data\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.248376 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2jwr\" (UniqueName: \"kubernetes.io/projected/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-kube-api-access-j2jwr\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.249557 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qp2fk\" (UniqueName: \"kubernetes.io/projected/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-kube-api-access-qp2fk\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.256231 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.259182 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-2\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.261218 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.280855 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-1\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.310880 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.340449 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.772456 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 14:08:21 crc kubenswrapper[4702]: W1125 14:08:21.781525 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbdf8d623_f3c0_4dc1_a034_f9a10c44ac39.slice/crio-4a3937b5a080bec6f9669a135f5af13451679a8c988e53b4b9cc506414b8d4bf WatchSource:0}: Error finding container 4a3937b5a080bec6f9669a135f5af13451679a8c988e53b4b9cc506414b8d4bf: Status 404 returned error can't find the container with id 4a3937b5a080bec6f9669a135f5af13451679a8c988e53b4b9cc506414b8d4bf Nov 25 14:08:21 crc kubenswrapper[4702]: I1125 14:08:21.841766 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 25 14:08:21 crc kubenswrapper[4702]: W1125 14:08:21.857481 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc2d6b57_c30a_4d66_a0d1_0cdc31e24309.slice/crio-bf9ae943a3b6a0805a5a90fe5efedb82e3e971c038597dd42bdf2d33d8639899 WatchSource:0}: Error finding container bf9ae943a3b6a0805a5a90fe5efedb82e3e971c038597dd42bdf2d33d8639899: Status 404 returned error can't find the container with id bf9ae943a3b6a0805a5a90fe5efedb82e3e971c038597dd42bdf2d33d8639899 Nov 25 14:08:22 crc kubenswrapper[4702]: I1125 14:08:22.098403 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39","Type":"ContainerStarted","Data":"bbc86ccc4f3ab91452eeb11ce870e3db074a7ef3732391144f5f800b1a5195b1"} Nov 25 14:08:22 crc kubenswrapper[4702]: I1125 14:08:22.098455 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39","Type":"ContainerStarted","Data":"4a3937b5a080bec6f9669a135f5af13451679a8c988e53b4b9cc506414b8d4bf"} Nov 25 14:08:22 crc kubenswrapper[4702]: I1125 14:08:22.101877 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309","Type":"ContainerStarted","Data":"9984e82fd4bc9b15be334e047fb511eba39e0e8c65b486ed771ea660e9fab2ed"} Nov 25 14:08:22 crc kubenswrapper[4702]: I1125 14:08:22.101914 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309","Type":"ContainerStarted","Data":"bf9ae943a3b6a0805a5a90fe5efedb82e3e971c038597dd42bdf2d33d8639899"} Nov 25 14:08:23 crc kubenswrapper[4702]: I1125 14:08:23.113270 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39","Type":"ContainerStarted","Data":"da18fb56c0ae01cd6deafd6e413cb4f839a5fe97179e4682ccc632db4f79afb9"} Nov 25 14:08:23 crc kubenswrapper[4702]: I1125 14:08:23.113898 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39","Type":"ContainerStarted","Data":"733ea752ee30c3751fb61d538f5869bde6516898dbd63fb278eb67b614d97715"} Nov 25 14:08:23 crc kubenswrapper[4702]: I1125 14:08:23.119013 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309","Type":"ContainerStarted","Data":"fb26ccc472afe01509779b387e91d8ccafe0951956cf4ff35c708aba13081b8a"} Nov 25 14:08:23 crc kubenswrapper[4702]: I1125 14:08:23.119064 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309","Type":"ContainerStarted","Data":"99811e41b3ab71c6d24ed9f0910b769fde45df4db75be46a99a49d5b0d52b925"} Nov 25 14:08:23 crc kubenswrapper[4702]: I1125 14:08:23.206775 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-1" podStartSLOduration=4.206754631 podStartE2EDuration="4.206754631s" podCreationTimestamp="2025-11-25 14:08:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 14:08:23.161055243 +0000 UTC m=+1312.323705703" watchObservedRunningTime="2025-11-25 14:08:23.206754631 +0000 UTC m=+1312.369405081" Nov 25 14:08:23 crc kubenswrapper[4702]: I1125 14:08:23.208814 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-2" podStartSLOduration=4.208805202 podStartE2EDuration="4.208805202s" podCreationTimestamp="2025-11-25 14:08:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 14:08:23.194487748 +0000 UTC m=+1312.357138198" watchObservedRunningTime="2025-11-25 14:08:23.208805202 +0000 UTC m=+1312.371455652" Nov 25 14:08:31 crc kubenswrapper[4702]: I1125 14:08:31.312290 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:31 crc kubenswrapper[4702]: I1125 14:08:31.312768 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:31 crc kubenswrapper[4702]: I1125 14:08:31.312780 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:31 crc kubenswrapper[4702]: I1125 14:08:31.339611 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:31 crc kubenswrapper[4702]: I1125 14:08:31.341576 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:31 crc kubenswrapper[4702]: I1125 14:08:31.342172 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:31 crc kubenswrapper[4702]: I1125 14:08:31.342213 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:31 crc kubenswrapper[4702]: I1125 14:08:31.346999 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:31 crc kubenswrapper[4702]: I1125 14:08:31.361729 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:31 crc kubenswrapper[4702]: I1125 14:08:31.372002 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:31 crc kubenswrapper[4702]: I1125 14:08:31.379391 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:31 crc kubenswrapper[4702]: I1125 14:08:31.387013 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:32 crc kubenswrapper[4702]: I1125 14:08:32.188335 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:32 crc kubenswrapper[4702]: I1125 14:08:32.188403 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:32 crc kubenswrapper[4702]: I1125 14:08:32.188416 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:32 crc kubenswrapper[4702]: I1125 14:08:32.188427 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:32 crc kubenswrapper[4702]: I1125 14:08:32.188437 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:32 crc kubenswrapper[4702]: I1125 14:08:32.188448 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:32 crc kubenswrapper[4702]: I1125 14:08:32.199857 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:32 crc kubenswrapper[4702]: I1125 14:08:32.200213 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:32 crc kubenswrapper[4702]: I1125 14:08:32.200746 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:32 crc kubenswrapper[4702]: I1125 14:08:32.200906 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:08:32 crc kubenswrapper[4702]: I1125 14:08:32.201523 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:08:32 crc kubenswrapper[4702]: I1125 14:08:32.205875 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:10:03 crc kubenswrapper[4702]: I1125 14:10:03.963472 4702 patch_prober.go:28] interesting pod/machine-config-daemon-qqr76 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 14:10:03 crc kubenswrapper[4702]: I1125 14:10:03.964145 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 14:10:33 crc kubenswrapper[4702]: I1125 14:10:33.963060 4702 patch_prober.go:28] interesting pod/machine-config-daemon-qqr76 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 14:10:33 crc kubenswrapper[4702]: I1125 14:10:33.963954 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 14:10:41 crc kubenswrapper[4702]: I1125 14:10:41.502196 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zsv8t"] Nov 25 14:10:41 crc kubenswrapper[4702]: I1125 14:10:41.504288 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zsv8t" Nov 25 14:10:41 crc kubenswrapper[4702]: I1125 14:10:41.542537 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zsv8t"] Nov 25 14:10:41 crc kubenswrapper[4702]: I1125 14:10:41.627444 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b834f31f-e712-4e98-aa8f-5d8a264bdbff-utilities\") pod \"certified-operators-zsv8t\" (UID: \"b834f31f-e712-4e98-aa8f-5d8a264bdbff\") " pod="openshift-marketplace/certified-operators-zsv8t" Nov 25 14:10:41 crc kubenswrapper[4702]: I1125 14:10:41.627592 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b834f31f-e712-4e98-aa8f-5d8a264bdbff-catalog-content\") pod \"certified-operators-zsv8t\" (UID: \"b834f31f-e712-4e98-aa8f-5d8a264bdbff\") " pod="openshift-marketplace/certified-operators-zsv8t" Nov 25 14:10:41 crc kubenswrapper[4702]: I1125 14:10:41.627828 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cj8hm\" (UniqueName: \"kubernetes.io/projected/b834f31f-e712-4e98-aa8f-5d8a264bdbff-kube-api-access-cj8hm\") pod \"certified-operators-zsv8t\" (UID: \"b834f31f-e712-4e98-aa8f-5d8a264bdbff\") " pod="openshift-marketplace/certified-operators-zsv8t" Nov 25 14:10:41 crc kubenswrapper[4702]: I1125 14:10:41.729721 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b834f31f-e712-4e98-aa8f-5d8a264bdbff-catalog-content\") pod \"certified-operators-zsv8t\" (UID: \"b834f31f-e712-4e98-aa8f-5d8a264bdbff\") " pod="openshift-marketplace/certified-operators-zsv8t" Nov 25 14:10:41 crc kubenswrapper[4702]: I1125 14:10:41.729897 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cj8hm\" (UniqueName: \"kubernetes.io/projected/b834f31f-e712-4e98-aa8f-5d8a264bdbff-kube-api-access-cj8hm\") pod \"certified-operators-zsv8t\" (UID: \"b834f31f-e712-4e98-aa8f-5d8a264bdbff\") " pod="openshift-marketplace/certified-operators-zsv8t" Nov 25 14:10:41 crc kubenswrapper[4702]: I1125 14:10:41.729978 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b834f31f-e712-4e98-aa8f-5d8a264bdbff-utilities\") pod \"certified-operators-zsv8t\" (UID: \"b834f31f-e712-4e98-aa8f-5d8a264bdbff\") " pod="openshift-marketplace/certified-operators-zsv8t" Nov 25 14:10:41 crc kubenswrapper[4702]: I1125 14:10:41.730352 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b834f31f-e712-4e98-aa8f-5d8a264bdbff-catalog-content\") pod \"certified-operators-zsv8t\" (UID: \"b834f31f-e712-4e98-aa8f-5d8a264bdbff\") " pod="openshift-marketplace/certified-operators-zsv8t" Nov 25 14:10:41 crc kubenswrapper[4702]: I1125 14:10:41.730602 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b834f31f-e712-4e98-aa8f-5d8a264bdbff-utilities\") pod \"certified-operators-zsv8t\" (UID: \"b834f31f-e712-4e98-aa8f-5d8a264bdbff\") " pod="openshift-marketplace/certified-operators-zsv8t" Nov 25 14:10:41 crc kubenswrapper[4702]: I1125 14:10:41.754207 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cj8hm\" (UniqueName: \"kubernetes.io/projected/b834f31f-e712-4e98-aa8f-5d8a264bdbff-kube-api-access-cj8hm\") pod \"certified-operators-zsv8t\" (UID: \"b834f31f-e712-4e98-aa8f-5d8a264bdbff\") " pod="openshift-marketplace/certified-operators-zsv8t" Nov 25 14:10:41 crc kubenswrapper[4702]: I1125 14:10:41.839481 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zsv8t" Nov 25 14:10:42 crc kubenswrapper[4702]: I1125 14:10:42.362532 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zsv8t"] Nov 25 14:10:43 crc kubenswrapper[4702]: I1125 14:10:43.277208 4702 generic.go:334] "Generic (PLEG): container finished" podID="b834f31f-e712-4e98-aa8f-5d8a264bdbff" containerID="bf30b948a9cff6cb0e88365542be0eeafabb16b723114157c3151402a6e833af" exitCode=0 Nov 25 14:10:43 crc kubenswrapper[4702]: I1125 14:10:43.277326 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zsv8t" event={"ID":"b834f31f-e712-4e98-aa8f-5d8a264bdbff","Type":"ContainerDied","Data":"bf30b948a9cff6cb0e88365542be0eeafabb16b723114157c3151402a6e833af"} Nov 25 14:10:43 crc kubenswrapper[4702]: I1125 14:10:43.277559 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zsv8t" event={"ID":"b834f31f-e712-4e98-aa8f-5d8a264bdbff","Type":"ContainerStarted","Data":"6e101e6de49618f50b47dac8ed61d2b7ad606f1491f6834fc712db45361ede1b"} Nov 25 14:10:43 crc kubenswrapper[4702]: I1125 14:10:43.280275 4702 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 14:10:45 crc kubenswrapper[4702]: I1125 14:10:45.299313 4702 generic.go:334] "Generic (PLEG): container finished" podID="b834f31f-e712-4e98-aa8f-5d8a264bdbff" containerID="83c0fb08f4a15bef5e89bba40ba73841ddbbfc9654a46c2e27c3f6f89c7c0b4f" exitCode=0 Nov 25 14:10:45 crc kubenswrapper[4702]: I1125 14:10:45.299473 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zsv8t" event={"ID":"b834f31f-e712-4e98-aa8f-5d8a264bdbff","Type":"ContainerDied","Data":"83c0fb08f4a15bef5e89bba40ba73841ddbbfc9654a46c2e27c3f6f89c7c0b4f"} Nov 25 14:10:46 crc kubenswrapper[4702]: I1125 14:10:46.315084 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zsv8t" event={"ID":"b834f31f-e712-4e98-aa8f-5d8a264bdbff","Type":"ContainerStarted","Data":"a78bc57a0b15e9a4310ba0a74b9c083c4bd1bda594431dd4aa3d3880d8aeebcf"} Nov 25 14:10:46 crc kubenswrapper[4702]: I1125 14:10:46.361913 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zsv8t" podStartSLOduration=2.748128748 podStartE2EDuration="5.361879096s" podCreationTimestamp="2025-11-25 14:10:41 +0000 UTC" firstStartedPulling="2025-11-25 14:10:43.279986911 +0000 UTC m=+1452.442637361" lastFinishedPulling="2025-11-25 14:10:45.893737269 +0000 UTC m=+1455.056387709" observedRunningTime="2025-11-25 14:10:46.343990416 +0000 UTC m=+1455.506640946" watchObservedRunningTime="2025-11-25 14:10:46.361879096 +0000 UTC m=+1455.524529586" Nov 25 14:10:51 crc kubenswrapper[4702]: I1125 14:10:51.840552 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zsv8t" Nov 25 14:10:51 crc kubenswrapper[4702]: I1125 14:10:51.841207 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zsv8t" Nov 25 14:10:51 crc kubenswrapper[4702]: I1125 14:10:51.887302 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zsv8t" Nov 25 14:10:52 crc kubenswrapper[4702]: I1125 14:10:52.418583 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zsv8t" Nov 25 14:10:52 crc kubenswrapper[4702]: I1125 14:10:52.477929 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zsv8t"] Nov 25 14:10:54 crc kubenswrapper[4702]: I1125 14:10:54.383113 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zsv8t" podUID="b834f31f-e712-4e98-aa8f-5d8a264bdbff" containerName="registry-server" containerID="cri-o://a78bc57a0b15e9a4310ba0a74b9c083c4bd1bda594431dd4aa3d3880d8aeebcf" gracePeriod=2 Nov 25 14:10:55 crc kubenswrapper[4702]: I1125 14:10:55.400080 4702 generic.go:334] "Generic (PLEG): container finished" podID="b834f31f-e712-4e98-aa8f-5d8a264bdbff" containerID="a78bc57a0b15e9a4310ba0a74b9c083c4bd1bda594431dd4aa3d3880d8aeebcf" exitCode=0 Nov 25 14:10:55 crc kubenswrapper[4702]: I1125 14:10:55.400153 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zsv8t" event={"ID":"b834f31f-e712-4e98-aa8f-5d8a264bdbff","Type":"ContainerDied","Data":"a78bc57a0b15e9a4310ba0a74b9c083c4bd1bda594431dd4aa3d3880d8aeebcf"} Nov 25 14:10:55 crc kubenswrapper[4702]: I1125 14:10:55.958009 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zsv8t" Nov 25 14:10:56 crc kubenswrapper[4702]: I1125 14:10:56.101925 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b834f31f-e712-4e98-aa8f-5d8a264bdbff-catalog-content\") pod \"b834f31f-e712-4e98-aa8f-5d8a264bdbff\" (UID: \"b834f31f-e712-4e98-aa8f-5d8a264bdbff\") " Nov 25 14:10:56 crc kubenswrapper[4702]: I1125 14:10:56.102141 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cj8hm\" (UniqueName: \"kubernetes.io/projected/b834f31f-e712-4e98-aa8f-5d8a264bdbff-kube-api-access-cj8hm\") pod \"b834f31f-e712-4e98-aa8f-5d8a264bdbff\" (UID: \"b834f31f-e712-4e98-aa8f-5d8a264bdbff\") " Nov 25 14:10:56 crc kubenswrapper[4702]: I1125 14:10:56.102220 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b834f31f-e712-4e98-aa8f-5d8a264bdbff-utilities\") pod \"b834f31f-e712-4e98-aa8f-5d8a264bdbff\" (UID: \"b834f31f-e712-4e98-aa8f-5d8a264bdbff\") " Nov 25 14:10:56 crc kubenswrapper[4702]: I1125 14:10:56.103254 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b834f31f-e712-4e98-aa8f-5d8a264bdbff-utilities" (OuterVolumeSpecName: "utilities") pod "b834f31f-e712-4e98-aa8f-5d8a264bdbff" (UID: "b834f31f-e712-4e98-aa8f-5d8a264bdbff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:10:56 crc kubenswrapper[4702]: I1125 14:10:56.116509 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b834f31f-e712-4e98-aa8f-5d8a264bdbff-kube-api-access-cj8hm" (OuterVolumeSpecName: "kube-api-access-cj8hm") pod "b834f31f-e712-4e98-aa8f-5d8a264bdbff" (UID: "b834f31f-e712-4e98-aa8f-5d8a264bdbff"). InnerVolumeSpecName "kube-api-access-cj8hm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:10:56 crc kubenswrapper[4702]: I1125 14:10:56.204579 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cj8hm\" (UniqueName: \"kubernetes.io/projected/b834f31f-e712-4e98-aa8f-5d8a264bdbff-kube-api-access-cj8hm\") on node \"crc\" DevicePath \"\"" Nov 25 14:10:56 crc kubenswrapper[4702]: I1125 14:10:56.204616 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b834f31f-e712-4e98-aa8f-5d8a264bdbff-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 14:10:56 crc kubenswrapper[4702]: I1125 14:10:56.414137 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zsv8t" event={"ID":"b834f31f-e712-4e98-aa8f-5d8a264bdbff","Type":"ContainerDied","Data":"6e101e6de49618f50b47dac8ed61d2b7ad606f1491f6834fc712db45361ede1b"} Nov 25 14:10:56 crc kubenswrapper[4702]: I1125 14:10:56.414203 4702 scope.go:117] "RemoveContainer" containerID="a78bc57a0b15e9a4310ba0a74b9c083c4bd1bda594431dd4aa3d3880d8aeebcf" Nov 25 14:10:56 crc kubenswrapper[4702]: I1125 14:10:56.414222 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zsv8t" Nov 25 14:10:56 crc kubenswrapper[4702]: I1125 14:10:56.432305 4702 scope.go:117] "RemoveContainer" containerID="83c0fb08f4a15bef5e89bba40ba73841ddbbfc9654a46c2e27c3f6f89c7c0b4f" Nov 25 14:10:56 crc kubenswrapper[4702]: I1125 14:10:56.451631 4702 scope.go:117] "RemoveContainer" containerID="bf30b948a9cff6cb0e88365542be0eeafabb16b723114157c3151402a6e833af" Nov 25 14:10:57 crc kubenswrapper[4702]: I1125 14:10:57.507951 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b834f31f-e712-4e98-aa8f-5d8a264bdbff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b834f31f-e712-4e98-aa8f-5d8a264bdbff" (UID: "b834f31f-e712-4e98-aa8f-5d8a264bdbff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:10:57 crc kubenswrapper[4702]: I1125 14:10:57.525596 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b834f31f-e712-4e98-aa8f-5d8a264bdbff-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 14:10:57 crc kubenswrapper[4702]: I1125 14:10:57.650728 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zsv8t"] Nov 25 14:10:57 crc kubenswrapper[4702]: I1125 14:10:57.657466 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zsv8t"] Nov 25 14:10:57 crc kubenswrapper[4702]: I1125 14:10:57.816943 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b834f31f-e712-4e98-aa8f-5d8a264bdbff" path="/var/lib/kubelet/pods/b834f31f-e712-4e98-aa8f-5d8a264bdbff/volumes" Nov 25 14:11:03 crc kubenswrapper[4702]: I1125 14:11:03.963326 4702 patch_prober.go:28] interesting pod/machine-config-daemon-qqr76 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 14:11:03 crc kubenswrapper[4702]: I1125 14:11:03.964038 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 14:11:03 crc kubenswrapper[4702]: I1125 14:11:03.964096 4702 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" Nov 25 14:11:03 crc kubenswrapper[4702]: I1125 14:11:03.964963 4702 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3c4335b16016506091cd0c40c0eeaec31932b0b8e710fd86970dfc2371661ab2"} pod="openshift-machine-config-operator/machine-config-daemon-qqr76" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 14:11:03 crc kubenswrapper[4702]: I1125 14:11:03.965038 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" containerID="cri-o://3c4335b16016506091cd0c40c0eeaec31932b0b8e710fd86970dfc2371661ab2" gracePeriod=600 Nov 25 14:11:04 crc kubenswrapper[4702]: I1125 14:11:04.488050 4702 generic.go:334] "Generic (PLEG): container finished" podID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerID="3c4335b16016506091cd0c40c0eeaec31932b0b8e710fd86970dfc2371661ab2" exitCode=0 Nov 25 14:11:04 crc kubenswrapper[4702]: I1125 14:11:04.488125 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" event={"ID":"3c491818-b0bb-4d82-a031-96e2dfac8c27","Type":"ContainerDied","Data":"3c4335b16016506091cd0c40c0eeaec31932b0b8e710fd86970dfc2371661ab2"} Nov 25 14:11:04 crc kubenswrapper[4702]: I1125 14:11:04.488564 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" event={"ID":"3c491818-b0bb-4d82-a031-96e2dfac8c27","Type":"ContainerStarted","Data":"32273717263d7fcc99e8193a439dd976da66869c55bb4633d9314cacb6ab92a1"} Nov 25 14:11:04 crc kubenswrapper[4702]: I1125 14:11:04.488584 4702 scope.go:117] "RemoveContainer" containerID="de4c184eb16dc995f1ebd5c48e0561cca0a21907587a6c388b21643c8a15afa4" Nov 25 14:11:18 crc kubenswrapper[4702]: I1125 14:11:18.772750 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dqgk4"] Nov 25 14:11:18 crc kubenswrapper[4702]: E1125 14:11:18.773561 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b834f31f-e712-4e98-aa8f-5d8a264bdbff" containerName="registry-server" Nov 25 14:11:18 crc kubenswrapper[4702]: I1125 14:11:18.773575 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="b834f31f-e712-4e98-aa8f-5d8a264bdbff" containerName="registry-server" Nov 25 14:11:18 crc kubenswrapper[4702]: E1125 14:11:18.773610 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b834f31f-e712-4e98-aa8f-5d8a264bdbff" containerName="extract-utilities" Nov 25 14:11:18 crc kubenswrapper[4702]: I1125 14:11:18.773616 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="b834f31f-e712-4e98-aa8f-5d8a264bdbff" containerName="extract-utilities" Nov 25 14:11:18 crc kubenswrapper[4702]: E1125 14:11:18.773631 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b834f31f-e712-4e98-aa8f-5d8a264bdbff" containerName="extract-content" Nov 25 14:11:18 crc kubenswrapper[4702]: I1125 14:11:18.773639 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="b834f31f-e712-4e98-aa8f-5d8a264bdbff" containerName="extract-content" Nov 25 14:11:18 crc kubenswrapper[4702]: I1125 14:11:18.773794 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="b834f31f-e712-4e98-aa8f-5d8a264bdbff" containerName="registry-server" Nov 25 14:11:18 crc kubenswrapper[4702]: I1125 14:11:18.774803 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dqgk4" Nov 25 14:11:18 crc kubenswrapper[4702]: I1125 14:11:18.791081 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dqgk4"] Nov 25 14:11:18 crc kubenswrapper[4702]: I1125 14:11:18.922841 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z96jl\" (UniqueName: \"kubernetes.io/projected/cecda5cc-95e9-4966-a1dd-702cb78e4623-kube-api-access-z96jl\") pod \"community-operators-dqgk4\" (UID: \"cecda5cc-95e9-4966-a1dd-702cb78e4623\") " pod="openshift-marketplace/community-operators-dqgk4" Nov 25 14:11:18 crc kubenswrapper[4702]: I1125 14:11:18.922967 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cecda5cc-95e9-4966-a1dd-702cb78e4623-utilities\") pod \"community-operators-dqgk4\" (UID: \"cecda5cc-95e9-4966-a1dd-702cb78e4623\") " pod="openshift-marketplace/community-operators-dqgk4" Nov 25 14:11:18 crc kubenswrapper[4702]: I1125 14:11:18.924149 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cecda5cc-95e9-4966-a1dd-702cb78e4623-catalog-content\") pod \"community-operators-dqgk4\" (UID: \"cecda5cc-95e9-4966-a1dd-702cb78e4623\") " pod="openshift-marketplace/community-operators-dqgk4" Nov 25 14:11:19 crc kubenswrapper[4702]: I1125 14:11:19.025960 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z96jl\" (UniqueName: \"kubernetes.io/projected/cecda5cc-95e9-4966-a1dd-702cb78e4623-kube-api-access-z96jl\") pod \"community-operators-dqgk4\" (UID: \"cecda5cc-95e9-4966-a1dd-702cb78e4623\") " pod="openshift-marketplace/community-operators-dqgk4" Nov 25 14:11:19 crc kubenswrapper[4702]: I1125 14:11:19.026043 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cecda5cc-95e9-4966-a1dd-702cb78e4623-utilities\") pod \"community-operators-dqgk4\" (UID: \"cecda5cc-95e9-4966-a1dd-702cb78e4623\") " pod="openshift-marketplace/community-operators-dqgk4" Nov 25 14:11:19 crc kubenswrapper[4702]: I1125 14:11:19.026159 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cecda5cc-95e9-4966-a1dd-702cb78e4623-catalog-content\") pod \"community-operators-dqgk4\" (UID: \"cecda5cc-95e9-4966-a1dd-702cb78e4623\") " pod="openshift-marketplace/community-operators-dqgk4" Nov 25 14:11:19 crc kubenswrapper[4702]: I1125 14:11:19.026805 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cecda5cc-95e9-4966-a1dd-702cb78e4623-catalog-content\") pod \"community-operators-dqgk4\" (UID: \"cecda5cc-95e9-4966-a1dd-702cb78e4623\") " pod="openshift-marketplace/community-operators-dqgk4" Nov 25 14:11:19 crc kubenswrapper[4702]: I1125 14:11:19.027011 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cecda5cc-95e9-4966-a1dd-702cb78e4623-utilities\") pod \"community-operators-dqgk4\" (UID: \"cecda5cc-95e9-4966-a1dd-702cb78e4623\") " pod="openshift-marketplace/community-operators-dqgk4" Nov 25 14:11:19 crc kubenswrapper[4702]: I1125 14:11:19.044721 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z96jl\" (UniqueName: \"kubernetes.io/projected/cecda5cc-95e9-4966-a1dd-702cb78e4623-kube-api-access-z96jl\") pod \"community-operators-dqgk4\" (UID: \"cecda5cc-95e9-4966-a1dd-702cb78e4623\") " pod="openshift-marketplace/community-operators-dqgk4" Nov 25 14:11:19 crc kubenswrapper[4702]: I1125 14:11:19.092968 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dqgk4" Nov 25 14:11:19 crc kubenswrapper[4702]: I1125 14:11:19.636063 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dqgk4"] Nov 25 14:11:20 crc kubenswrapper[4702]: I1125 14:11:20.625620 4702 generic.go:334] "Generic (PLEG): container finished" podID="cecda5cc-95e9-4966-a1dd-702cb78e4623" containerID="a2dff3a60d75e2f4ec83714498be92f9f7132b1ee078e6e2ae4cd347c2912723" exitCode=0 Nov 25 14:11:20 crc kubenswrapper[4702]: I1125 14:11:20.625907 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dqgk4" event={"ID":"cecda5cc-95e9-4966-a1dd-702cb78e4623","Type":"ContainerDied","Data":"a2dff3a60d75e2f4ec83714498be92f9f7132b1ee078e6e2ae4cd347c2912723"} Nov 25 14:11:20 crc kubenswrapper[4702]: I1125 14:11:20.625939 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dqgk4" event={"ID":"cecda5cc-95e9-4966-a1dd-702cb78e4623","Type":"ContainerStarted","Data":"8739ed3f3e152911818c27c429ef114b93430c1c6717ac139f8450a828b97e7d"} Nov 25 14:11:21 crc kubenswrapper[4702]: I1125 14:11:21.634918 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dqgk4" event={"ID":"cecda5cc-95e9-4966-a1dd-702cb78e4623","Type":"ContainerStarted","Data":"67598927e837c4f65f899d883d79bf35919e06156daa0ea556b002c10057308e"} Nov 25 14:11:22 crc kubenswrapper[4702]: I1125 14:11:22.643460 4702 generic.go:334] "Generic (PLEG): container finished" podID="cecda5cc-95e9-4966-a1dd-702cb78e4623" containerID="67598927e837c4f65f899d883d79bf35919e06156daa0ea556b002c10057308e" exitCode=0 Nov 25 14:11:22 crc kubenswrapper[4702]: I1125 14:11:22.643529 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dqgk4" event={"ID":"cecda5cc-95e9-4966-a1dd-702cb78e4623","Type":"ContainerDied","Data":"67598927e837c4f65f899d883d79bf35919e06156daa0ea556b002c10057308e"} Nov 25 14:11:22 crc kubenswrapper[4702]: I1125 14:11:22.954438 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 25 14:11:22 crc kubenswrapper[4702]: I1125 14:11:22.955106 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-2" podUID="bc2d6b57-c30a-4d66-a0d1-0cdc31e24309" containerName="glance-log" containerID="cri-o://9984e82fd4bc9b15be334e047fb511eba39e0e8c65b486ed771ea660e9fab2ed" gracePeriod=30 Nov 25 14:11:22 crc kubenswrapper[4702]: I1125 14:11:22.955525 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-2" podUID="bc2d6b57-c30a-4d66-a0d1-0cdc31e24309" containerName="glance-api" containerID="cri-o://fb26ccc472afe01509779b387e91d8ccafe0951956cf4ff35c708aba13081b8a" gracePeriod=30 Nov 25 14:11:22 crc kubenswrapper[4702]: I1125 14:11:22.955572 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-2" podUID="bc2d6b57-c30a-4d66-a0d1-0cdc31e24309" containerName="glance-httpd" containerID="cri-o://99811e41b3ab71c6d24ed9f0910b769fde45df4db75be46a99a49d5b0d52b925" gracePeriod=30 Nov 25 14:11:22 crc kubenswrapper[4702]: I1125 14:11:22.968302 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 14:11:22 crc kubenswrapper[4702]: I1125 14:11:22.968600 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="bdf8d623-f3c0-4dc1-a034-f9a10c44ac39" containerName="glance-log" containerID="cri-o://bbc86ccc4f3ab91452eeb11ce870e3db074a7ef3732391144f5f800b1a5195b1" gracePeriod=30 Nov 25 14:11:22 crc kubenswrapper[4702]: I1125 14:11:22.968770 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="bdf8d623-f3c0-4dc1-a034-f9a10c44ac39" containerName="glance-api" containerID="cri-o://da18fb56c0ae01cd6deafd6e413cb4f839a5fe97179e4682ccc632db4f79afb9" gracePeriod=30 Nov 25 14:11:22 crc kubenswrapper[4702]: I1125 14:11:22.968810 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="bdf8d623-f3c0-4dc1-a034-f9a10c44ac39" containerName="glance-httpd" containerID="cri-o://733ea752ee30c3751fb61d538f5869bde6516898dbd63fb278eb67b614d97715" gracePeriod=30 Nov 25 14:11:23 crc kubenswrapper[4702]: I1125 14:11:23.653632 4702 generic.go:334] "Generic (PLEG): container finished" podID="bdf8d623-f3c0-4dc1-a034-f9a10c44ac39" containerID="da18fb56c0ae01cd6deafd6e413cb4f839a5fe97179e4682ccc632db4f79afb9" exitCode=0 Nov 25 14:11:23 crc kubenswrapper[4702]: I1125 14:11:23.653994 4702 generic.go:334] "Generic (PLEG): container finished" podID="bdf8d623-f3c0-4dc1-a034-f9a10c44ac39" containerID="733ea752ee30c3751fb61d538f5869bde6516898dbd63fb278eb67b614d97715" exitCode=0 Nov 25 14:11:23 crc kubenswrapper[4702]: I1125 14:11:23.654007 4702 generic.go:334] "Generic (PLEG): container finished" podID="bdf8d623-f3c0-4dc1-a034-f9a10c44ac39" containerID="bbc86ccc4f3ab91452eeb11ce870e3db074a7ef3732391144f5f800b1a5195b1" exitCode=143 Nov 25 14:11:23 crc kubenswrapper[4702]: I1125 14:11:23.653703 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39","Type":"ContainerDied","Data":"da18fb56c0ae01cd6deafd6e413cb4f839a5fe97179e4682ccc632db4f79afb9"} Nov 25 14:11:23 crc kubenswrapper[4702]: I1125 14:11:23.654088 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39","Type":"ContainerDied","Data":"733ea752ee30c3751fb61d538f5869bde6516898dbd63fb278eb67b614d97715"} Nov 25 14:11:23 crc kubenswrapper[4702]: I1125 14:11:23.654109 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39","Type":"ContainerDied","Data":"bbc86ccc4f3ab91452eeb11ce870e3db074a7ef3732391144f5f800b1a5195b1"} Nov 25 14:11:23 crc kubenswrapper[4702]: I1125 14:11:23.656834 4702 generic.go:334] "Generic (PLEG): container finished" podID="bc2d6b57-c30a-4d66-a0d1-0cdc31e24309" containerID="fb26ccc472afe01509779b387e91d8ccafe0951956cf4ff35c708aba13081b8a" exitCode=0 Nov 25 14:11:23 crc kubenswrapper[4702]: I1125 14:11:23.656864 4702 generic.go:334] "Generic (PLEG): container finished" podID="bc2d6b57-c30a-4d66-a0d1-0cdc31e24309" containerID="99811e41b3ab71c6d24ed9f0910b769fde45df4db75be46a99a49d5b0d52b925" exitCode=0 Nov 25 14:11:23 crc kubenswrapper[4702]: I1125 14:11:23.656871 4702 generic.go:334] "Generic (PLEG): container finished" podID="bc2d6b57-c30a-4d66-a0d1-0cdc31e24309" containerID="9984e82fd4bc9b15be334e047fb511eba39e0e8c65b486ed771ea660e9fab2ed" exitCode=143 Nov 25 14:11:23 crc kubenswrapper[4702]: I1125 14:11:23.656917 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309","Type":"ContainerDied","Data":"fb26ccc472afe01509779b387e91d8ccafe0951956cf4ff35c708aba13081b8a"} Nov 25 14:11:23 crc kubenswrapper[4702]: I1125 14:11:23.656981 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309","Type":"ContainerDied","Data":"99811e41b3ab71c6d24ed9f0910b769fde45df4db75be46a99a49d5b0d52b925"} Nov 25 14:11:23 crc kubenswrapper[4702]: I1125 14:11:23.656994 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309","Type":"ContainerDied","Data":"9984e82fd4bc9b15be334e047fb511eba39e0e8c65b486ed771ea660e9fab2ed"} Nov 25 14:11:23 crc kubenswrapper[4702]: I1125 14:11:23.660882 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dqgk4" event={"ID":"cecda5cc-95e9-4966-a1dd-702cb78e4623","Type":"ContainerStarted","Data":"c6a84396dbcaa8e9796180d0d842af7b05bae72486b7c07eae992e3ac3f78bb5"} Nov 25 14:11:23 crc kubenswrapper[4702]: I1125 14:11:23.685422 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dqgk4" podStartSLOduration=3.263245545 podStartE2EDuration="5.685398725s" podCreationTimestamp="2025-11-25 14:11:18 +0000 UTC" firstStartedPulling="2025-11-25 14:11:20.628082628 +0000 UTC m=+1489.790733078" lastFinishedPulling="2025-11-25 14:11:23.050235808 +0000 UTC m=+1492.212886258" observedRunningTime="2025-11-25 14:11:23.679817546 +0000 UTC m=+1492.842467996" watchObservedRunningTime="2025-11-25 14:11:23.685398725 +0000 UTC m=+1492.848049175" Nov 25 14:11:23 crc kubenswrapper[4702]: I1125 14:11:23.951140 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.019378 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.124148 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-run\") pod \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.124200 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-run\") pod \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.124222 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-sys\") pod \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.124307 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-config-data\") pod \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.124330 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-etc-iscsi\") pod \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.124364 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.124408 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-var-locks-brick\") pod \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.124439 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-var-locks-brick\") pod \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.124482 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2jwr\" (UniqueName: \"kubernetes.io/projected/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-kube-api-access-j2jwr\") pod \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.124506 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.124525 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.124546 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-etc-nvme\") pod \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.124566 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-etc-iscsi\") pod \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.124594 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-scripts\") pod \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.124627 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.124658 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-dev\") pod \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.124685 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-httpd-run\") pod \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.124729 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-dev\") pod \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.125997 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-config-data\") pod \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.126331 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-scripts\") pod \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.126355 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-sys\") pod \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.126378 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-lib-modules\") pod \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.126420 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-httpd-run\") pod \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.126449 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-logs\") pod \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.126477 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-etc-nvme\") pod \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.126505 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qp2fk\" (UniqueName: \"kubernetes.io/projected/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-kube-api-access-qp2fk\") pod \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\" (UID: \"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39\") " Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.126527 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-lib-modules\") pod \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.126552 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-logs\") pod \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\" (UID: \"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309\") " Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.124723 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "bdf8d623-f3c0-4dc1-a034-f9a10c44ac39" (UID: "bdf8d623-f3c0-4dc1-a034-f9a10c44ac39"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.124780 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-run" (OuterVolumeSpecName: "run") pod "bc2d6b57-c30a-4d66-a0d1-0cdc31e24309" (UID: "bc2d6b57-c30a-4d66-a0d1-0cdc31e24309"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.124801 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-run" (OuterVolumeSpecName: "run") pod "bdf8d623-f3c0-4dc1-a034-f9a10c44ac39" (UID: "bdf8d623-f3c0-4dc1-a034-f9a10c44ac39"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.124843 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-sys" (OuterVolumeSpecName: "sys") pod "bdf8d623-f3c0-4dc1-a034-f9a10c44ac39" (UID: "bdf8d623-f3c0-4dc1-a034-f9a10c44ac39"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.125464 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "bc2d6b57-c30a-4d66-a0d1-0cdc31e24309" (UID: "bc2d6b57-c30a-4d66-a0d1-0cdc31e24309"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.125487 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "bc2d6b57-c30a-4d66-a0d1-0cdc31e24309" (UID: "bc2d6b57-c30a-4d66-a0d1-0cdc31e24309"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.127527 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-dev" (OuterVolumeSpecName: "dev") pod "bc2d6b57-c30a-4d66-a0d1-0cdc31e24309" (UID: "bc2d6b57-c30a-4d66-a0d1-0cdc31e24309"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.127567 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "bc2d6b57-c30a-4d66-a0d1-0cdc31e24309" (UID: "bc2d6b57-c30a-4d66-a0d1-0cdc31e24309"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.127581 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-logs" (OuterVolumeSpecName: "logs") pod "bc2d6b57-c30a-4d66-a0d1-0cdc31e24309" (UID: "bc2d6b57-c30a-4d66-a0d1-0cdc31e24309"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.127716 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "bdf8d623-f3c0-4dc1-a034-f9a10c44ac39" (UID: "bdf8d623-f3c0-4dc1-a034-f9a10c44ac39"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.128183 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "bc2d6b57-c30a-4d66-a0d1-0cdc31e24309" (UID: "bc2d6b57-c30a-4d66-a0d1-0cdc31e24309"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.128244 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-logs" (OuterVolumeSpecName: "logs") pod "bdf8d623-f3c0-4dc1-a034-f9a10c44ac39" (UID: "bdf8d623-f3c0-4dc1-a034-f9a10c44ac39"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.128274 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-sys" (OuterVolumeSpecName: "sys") pod "bc2d6b57-c30a-4d66-a0d1-0cdc31e24309" (UID: "bc2d6b57-c30a-4d66-a0d1-0cdc31e24309"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.128303 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "bdf8d623-f3c0-4dc1-a034-f9a10c44ac39" (UID: "bdf8d623-f3c0-4dc1-a034-f9a10c44ac39"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.128532 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "bdf8d623-f3c0-4dc1-a034-f9a10c44ac39" (UID: "bdf8d623-f3c0-4dc1-a034-f9a10c44ac39"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.129412 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-dev" (OuterVolumeSpecName: "dev") pod "bdf8d623-f3c0-4dc1-a034-f9a10c44ac39" (UID: "bdf8d623-f3c0-4dc1-a034-f9a10c44ac39"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.129621 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "bdf8d623-f3c0-4dc1-a034-f9a10c44ac39" (UID: "bdf8d623-f3c0-4dc1-a034-f9a10c44ac39"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.129655 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "bc2d6b57-c30a-4d66-a0d1-0cdc31e24309" (UID: "bc2d6b57-c30a-4d66-a0d1-0cdc31e24309"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.138121 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-kube-api-access-j2jwr" (OuterVolumeSpecName: "kube-api-access-j2jwr") pod "bc2d6b57-c30a-4d66-a0d1-0cdc31e24309" (UID: "bc2d6b57-c30a-4d66-a0d1-0cdc31e24309"). InnerVolumeSpecName "kube-api-access-j2jwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.140066 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "bdf8d623-f3c0-4dc1-a034-f9a10c44ac39" (UID: "bdf8d623-f3c0-4dc1-a034-f9a10c44ac39"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.140468 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-scripts" (OuterVolumeSpecName: "scripts") pod "bc2d6b57-c30a-4d66-a0d1-0cdc31e24309" (UID: "bc2d6b57-c30a-4d66-a0d1-0cdc31e24309"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.140647 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance-cache") pod "bdf8d623-f3c0-4dc1-a034-f9a10c44ac39" (UID: "bdf8d623-f3c0-4dc1-a034-f9a10c44ac39"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.142768 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-scripts" (OuterVolumeSpecName: "scripts") pod "bdf8d623-f3c0-4dc1-a034-f9a10c44ac39" (UID: "bdf8d623-f3c0-4dc1-a034-f9a10c44ac39"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.159570 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-kube-api-access-qp2fk" (OuterVolumeSpecName: "kube-api-access-qp2fk") pod "bdf8d623-f3c0-4dc1-a034-f9a10c44ac39" (UID: "bdf8d623-f3c0-4dc1-a034-f9a10c44ac39"). InnerVolumeSpecName "kube-api-access-qp2fk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.160733 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance-cache") pod "bc2d6b57-c30a-4d66-a0d1-0cdc31e24309" (UID: "bc2d6b57-c30a-4d66-a0d1-0cdc31e24309"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.166103 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "bc2d6b57-c30a-4d66-a0d1-0cdc31e24309" (UID: "bc2d6b57-c30a-4d66-a0d1-0cdc31e24309"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.217526 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-config-data" (OuterVolumeSpecName: "config-data") pod "bc2d6b57-c30a-4d66-a0d1-0cdc31e24309" (UID: "bc2d6b57-c30a-4d66-a0d1-0cdc31e24309"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.228437 4702 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-run\") on node \"crc\" DevicePath \"\"" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.228484 4702 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-sys\") on node \"crc\" DevicePath \"\"" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.228518 4702 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-run\") on node \"crc\" DevicePath \"\"" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.228529 4702 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.228566 4702 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.228580 4702 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.228592 4702 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.228604 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2jwr\" (UniqueName: \"kubernetes.io/projected/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-kube-api-access-j2jwr\") on node \"crc\" DevicePath \"\"" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.228621 4702 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.228636 4702 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.228648 4702 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.228659 4702 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.228673 4702 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.228743 4702 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.228756 4702 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-dev\") on node \"crc\" DevicePath \"\"" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.228767 4702 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.228779 4702 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-dev\") on node \"crc\" DevicePath \"\"" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.228791 4702 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.228804 4702 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.228814 4702 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-sys\") on node \"crc\" DevicePath \"\"" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.228826 4702 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.228837 4702 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.228848 4702 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-logs\") on node \"crc\" DevicePath \"\"" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.228859 4702 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.228871 4702 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.228882 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qp2fk\" (UniqueName: \"kubernetes.io/projected/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-kube-api-access-qp2fk\") on node \"crc\" DevicePath \"\"" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.228895 4702 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309-logs\") on node \"crc\" DevicePath \"\"" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.255578 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-config-data" (OuterVolumeSpecName: "config-data") pod "bdf8d623-f3c0-4dc1-a034-f9a10c44ac39" (UID: "bdf8d623-f3c0-4dc1-a034-f9a10c44ac39"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.265105 4702 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.267194 4702 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.270378 4702 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.273882 4702 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.330373 4702 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.330423 4702 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.330437 4702 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.330448 4702 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.330460 4702 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.672953 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"bdf8d623-f3c0-4dc1-a034-f9a10c44ac39","Type":"ContainerDied","Data":"4a3937b5a080bec6f9669a135f5af13451679a8c988e53b4b9cc506414b8d4bf"} Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.673022 4702 scope.go:117] "RemoveContainer" containerID="da18fb56c0ae01cd6deafd6e413cb4f839a5fe97179e4682ccc632db4f79afb9" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.673418 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.683590 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"bc2d6b57-c30a-4d66-a0d1-0cdc31e24309","Type":"ContainerDied","Data":"bf9ae943a3b6a0805a5a90fe5efedb82e3e971c038597dd42bdf2d33d8639899"} Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.683636 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.707449 4702 scope.go:117] "RemoveContainer" containerID="733ea752ee30c3751fb61d538f5869bde6516898dbd63fb278eb67b614d97715" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.721438 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.736290 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.744387 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.749889 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.754814 4702 scope.go:117] "RemoveContainer" containerID="bbc86ccc4f3ab91452eeb11ce870e3db074a7ef3732391144f5f800b1a5195b1" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.780616 4702 scope.go:117] "RemoveContainer" containerID="fb26ccc472afe01509779b387e91d8ccafe0951956cf4ff35c708aba13081b8a" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.805542 4702 scope.go:117] "RemoveContainer" containerID="99811e41b3ab71c6d24ed9f0910b769fde45df4db75be46a99a49d5b0d52b925" Nov 25 14:11:24 crc kubenswrapper[4702]: I1125 14:11:24.828730 4702 scope.go:117] "RemoveContainer" containerID="9984e82fd4bc9b15be334e047fb511eba39e0e8c65b486ed771ea660e9fab2ed" Nov 25 14:11:25 crc kubenswrapper[4702]: I1125 14:11:25.814645 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc2d6b57-c30a-4d66-a0d1-0cdc31e24309" path="/var/lib/kubelet/pods/bc2d6b57-c30a-4d66-a0d1-0cdc31e24309/volumes" Nov 25 14:11:25 crc kubenswrapper[4702]: I1125 14:11:25.815398 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdf8d623-f3c0-4dc1-a034-f9a10c44ac39" path="/var/lib/kubelet/pods/bdf8d623-f3c0-4dc1-a034-f9a10c44ac39/volumes" Nov 25 14:11:29 crc kubenswrapper[4702]: I1125 14:11:29.094622 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dqgk4" Nov 25 14:11:29 crc kubenswrapper[4702]: I1125 14:11:29.095591 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dqgk4" Nov 25 14:11:29 crc kubenswrapper[4702]: I1125 14:11:29.159315 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dqgk4" Nov 25 14:11:29 crc kubenswrapper[4702]: I1125 14:11:29.786488 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dqgk4" Nov 25 14:11:29 crc kubenswrapper[4702]: I1125 14:11:29.833672 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dqgk4"] Nov 25 14:11:31 crc kubenswrapper[4702]: I1125 14:11:31.747818 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dqgk4" podUID="cecda5cc-95e9-4966-a1dd-702cb78e4623" containerName="registry-server" containerID="cri-o://c6a84396dbcaa8e9796180d0d842af7b05bae72486b7c07eae992e3ac3f78bb5" gracePeriod=2 Nov 25 14:11:32 crc kubenswrapper[4702]: I1125 14:11:32.229695 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dqgk4" Nov 25 14:11:32 crc kubenswrapper[4702]: I1125 14:11:32.358688 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z96jl\" (UniqueName: \"kubernetes.io/projected/cecda5cc-95e9-4966-a1dd-702cb78e4623-kube-api-access-z96jl\") pod \"cecda5cc-95e9-4966-a1dd-702cb78e4623\" (UID: \"cecda5cc-95e9-4966-a1dd-702cb78e4623\") " Nov 25 14:11:32 crc kubenswrapper[4702]: I1125 14:11:32.358831 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cecda5cc-95e9-4966-a1dd-702cb78e4623-utilities\") pod \"cecda5cc-95e9-4966-a1dd-702cb78e4623\" (UID: \"cecda5cc-95e9-4966-a1dd-702cb78e4623\") " Nov 25 14:11:32 crc kubenswrapper[4702]: I1125 14:11:32.358934 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cecda5cc-95e9-4966-a1dd-702cb78e4623-catalog-content\") pod \"cecda5cc-95e9-4966-a1dd-702cb78e4623\" (UID: \"cecda5cc-95e9-4966-a1dd-702cb78e4623\") " Nov 25 14:11:32 crc kubenswrapper[4702]: I1125 14:11:32.362476 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cecda5cc-95e9-4966-a1dd-702cb78e4623-utilities" (OuterVolumeSpecName: "utilities") pod "cecda5cc-95e9-4966-a1dd-702cb78e4623" (UID: "cecda5cc-95e9-4966-a1dd-702cb78e4623"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:11:32 crc kubenswrapper[4702]: I1125 14:11:32.395438 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cecda5cc-95e9-4966-a1dd-702cb78e4623-kube-api-access-z96jl" (OuterVolumeSpecName: "kube-api-access-z96jl") pod "cecda5cc-95e9-4966-a1dd-702cb78e4623" (UID: "cecda5cc-95e9-4966-a1dd-702cb78e4623"). InnerVolumeSpecName "kube-api-access-z96jl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:11:32 crc kubenswrapper[4702]: I1125 14:11:32.461218 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cecda5cc-95e9-4966-a1dd-702cb78e4623-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 14:11:32 crc kubenswrapper[4702]: I1125 14:11:32.461278 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z96jl\" (UniqueName: \"kubernetes.io/projected/cecda5cc-95e9-4966-a1dd-702cb78e4623-kube-api-access-z96jl\") on node \"crc\" DevicePath \"\"" Nov 25 14:11:32 crc kubenswrapper[4702]: I1125 14:11:32.758421 4702 generic.go:334] "Generic (PLEG): container finished" podID="cecda5cc-95e9-4966-a1dd-702cb78e4623" containerID="c6a84396dbcaa8e9796180d0d842af7b05bae72486b7c07eae992e3ac3f78bb5" exitCode=0 Nov 25 14:11:32 crc kubenswrapper[4702]: I1125 14:11:32.758464 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dqgk4" Nov 25 14:11:32 crc kubenswrapper[4702]: I1125 14:11:32.758472 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dqgk4" event={"ID":"cecda5cc-95e9-4966-a1dd-702cb78e4623","Type":"ContainerDied","Data":"c6a84396dbcaa8e9796180d0d842af7b05bae72486b7c07eae992e3ac3f78bb5"} Nov 25 14:11:32 crc kubenswrapper[4702]: I1125 14:11:32.758523 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dqgk4" event={"ID":"cecda5cc-95e9-4966-a1dd-702cb78e4623","Type":"ContainerDied","Data":"8739ed3f3e152911818c27c429ef114b93430c1c6717ac139f8450a828b97e7d"} Nov 25 14:11:32 crc kubenswrapper[4702]: I1125 14:11:32.758542 4702 scope.go:117] "RemoveContainer" containerID="c6a84396dbcaa8e9796180d0d842af7b05bae72486b7c07eae992e3ac3f78bb5" Nov 25 14:11:32 crc kubenswrapper[4702]: I1125 14:11:32.795675 4702 scope.go:117] "RemoveContainer" containerID="67598927e837c4f65f899d883d79bf35919e06156daa0ea556b002c10057308e" Nov 25 14:11:32 crc kubenswrapper[4702]: I1125 14:11:32.824091 4702 scope.go:117] "RemoveContainer" containerID="a2dff3a60d75e2f4ec83714498be92f9f7132b1ee078e6e2ae4cd347c2912723" Nov 25 14:11:32 crc kubenswrapper[4702]: I1125 14:11:32.859931 4702 scope.go:117] "RemoveContainer" containerID="c6a84396dbcaa8e9796180d0d842af7b05bae72486b7c07eae992e3ac3f78bb5" Nov 25 14:11:32 crc kubenswrapper[4702]: E1125 14:11:32.860815 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6a84396dbcaa8e9796180d0d842af7b05bae72486b7c07eae992e3ac3f78bb5\": container with ID starting with c6a84396dbcaa8e9796180d0d842af7b05bae72486b7c07eae992e3ac3f78bb5 not found: ID does not exist" containerID="c6a84396dbcaa8e9796180d0d842af7b05bae72486b7c07eae992e3ac3f78bb5" Nov 25 14:11:32 crc kubenswrapper[4702]: I1125 14:11:32.860856 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6a84396dbcaa8e9796180d0d842af7b05bae72486b7c07eae992e3ac3f78bb5"} err="failed to get container status \"c6a84396dbcaa8e9796180d0d842af7b05bae72486b7c07eae992e3ac3f78bb5\": rpc error: code = NotFound desc = could not find container \"c6a84396dbcaa8e9796180d0d842af7b05bae72486b7c07eae992e3ac3f78bb5\": container with ID starting with c6a84396dbcaa8e9796180d0d842af7b05bae72486b7c07eae992e3ac3f78bb5 not found: ID does not exist" Nov 25 14:11:32 crc kubenswrapper[4702]: I1125 14:11:32.860900 4702 scope.go:117] "RemoveContainer" containerID="67598927e837c4f65f899d883d79bf35919e06156daa0ea556b002c10057308e" Nov 25 14:11:32 crc kubenswrapper[4702]: E1125 14:11:32.861343 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67598927e837c4f65f899d883d79bf35919e06156daa0ea556b002c10057308e\": container with ID starting with 67598927e837c4f65f899d883d79bf35919e06156daa0ea556b002c10057308e not found: ID does not exist" containerID="67598927e837c4f65f899d883d79bf35919e06156daa0ea556b002c10057308e" Nov 25 14:11:32 crc kubenswrapper[4702]: I1125 14:11:32.861400 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67598927e837c4f65f899d883d79bf35919e06156daa0ea556b002c10057308e"} err="failed to get container status \"67598927e837c4f65f899d883d79bf35919e06156daa0ea556b002c10057308e\": rpc error: code = NotFound desc = could not find container \"67598927e837c4f65f899d883d79bf35919e06156daa0ea556b002c10057308e\": container with ID starting with 67598927e837c4f65f899d883d79bf35919e06156daa0ea556b002c10057308e not found: ID does not exist" Nov 25 14:11:32 crc kubenswrapper[4702]: I1125 14:11:32.861424 4702 scope.go:117] "RemoveContainer" containerID="a2dff3a60d75e2f4ec83714498be92f9f7132b1ee078e6e2ae4cd347c2912723" Nov 25 14:11:32 crc kubenswrapper[4702]: E1125 14:11:32.861701 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2dff3a60d75e2f4ec83714498be92f9f7132b1ee078e6e2ae4cd347c2912723\": container with ID starting with a2dff3a60d75e2f4ec83714498be92f9f7132b1ee078e6e2ae4cd347c2912723 not found: ID does not exist" containerID="a2dff3a60d75e2f4ec83714498be92f9f7132b1ee078e6e2ae4cd347c2912723" Nov 25 14:11:32 crc kubenswrapper[4702]: I1125 14:11:32.861745 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2dff3a60d75e2f4ec83714498be92f9f7132b1ee078e6e2ae4cd347c2912723"} err="failed to get container status \"a2dff3a60d75e2f4ec83714498be92f9f7132b1ee078e6e2ae4cd347c2912723\": rpc error: code = NotFound desc = could not find container \"a2dff3a60d75e2f4ec83714498be92f9f7132b1ee078e6e2ae4cd347c2912723\": container with ID starting with a2dff3a60d75e2f4ec83714498be92f9f7132b1ee078e6e2ae4cd347c2912723 not found: ID does not exist" Nov 25 14:11:32 crc kubenswrapper[4702]: I1125 14:11:32.885087 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cecda5cc-95e9-4966-a1dd-702cb78e4623-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cecda5cc-95e9-4966-a1dd-702cb78e4623" (UID: "cecda5cc-95e9-4966-a1dd-702cb78e4623"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:11:32 crc kubenswrapper[4702]: I1125 14:11:32.957922 4702 scope.go:117] "RemoveContainer" containerID="4d9314c9f8c5153ba1f6da19cd33886fd23f8fdd2b8bfa73337a7a733a20d7d7" Nov 25 14:11:32 crc kubenswrapper[4702]: I1125 14:11:32.974619 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cecda5cc-95e9-4966-a1dd-702cb78e4623-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 14:11:33 crc kubenswrapper[4702]: I1125 14:11:33.008926 4702 scope.go:117] "RemoveContainer" containerID="c31eba3d0bcd450abc6b9d57105bb57edb5e7bb354f6e1d085bb6020923360ec" Nov 25 14:11:33 crc kubenswrapper[4702]: I1125 14:11:33.029035 4702 scope.go:117] "RemoveContainer" containerID="e1700aa0c7a2395cbe33b28a57ba9b6714d59955a966f02ff3b018bdf4db6635" Nov 25 14:11:33 crc kubenswrapper[4702]: I1125 14:11:33.107090 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dqgk4"] Nov 25 14:11:33 crc kubenswrapper[4702]: I1125 14:11:33.114547 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dqgk4"] Nov 25 14:11:33 crc kubenswrapper[4702]: I1125 14:11:33.810413 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cecda5cc-95e9-4966-a1dd-702cb78e4623" path="/var/lib/kubelet/pods/cecda5cc-95e9-4966-a1dd-702cb78e4623/volumes" Nov 25 14:11:34 crc kubenswrapper[4702]: E1125 14:11:34.331956 4702 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcecda5cc_95e9_4966_a1dd_702cb78e4623.slice/crio-c6a84396dbcaa8e9796180d0d842af7b05bae72486b7c07eae992e3ac3f78bb5.scope\": RecentStats: unable to find data in memory cache]" Nov 25 14:11:44 crc kubenswrapper[4702]: E1125 14:11:44.554746 4702 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcecda5cc_95e9_4966_a1dd_702cb78e4623.slice/crio-c6a84396dbcaa8e9796180d0d842af7b05bae72486b7c07eae992e3ac3f78bb5.scope\": RecentStats: unable to find data in memory cache]" Nov 25 14:11:54 crc kubenswrapper[4702]: E1125 14:11:54.801006 4702 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcecda5cc_95e9_4966_a1dd_702cb78e4623.slice/crio-c6a84396dbcaa8e9796180d0d842af7b05bae72486b7c07eae992e3ac3f78bb5.scope\": RecentStats: unable to find data in memory cache]" Nov 25 14:11:59 crc kubenswrapper[4702]: I1125 14:11:59.844679 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2sz5r"] Nov 25 14:11:59 crc kubenswrapper[4702]: E1125 14:11:59.849088 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdf8d623-f3c0-4dc1-a034-f9a10c44ac39" containerName="glance-api" Nov 25 14:11:59 crc kubenswrapper[4702]: I1125 14:11:59.849221 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdf8d623-f3c0-4dc1-a034-f9a10c44ac39" containerName="glance-api" Nov 25 14:11:59 crc kubenswrapper[4702]: E1125 14:11:59.849387 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cecda5cc-95e9-4966-a1dd-702cb78e4623" containerName="extract-utilities" Nov 25 14:11:59 crc kubenswrapper[4702]: I1125 14:11:59.849474 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="cecda5cc-95e9-4966-a1dd-702cb78e4623" containerName="extract-utilities" Nov 25 14:11:59 crc kubenswrapper[4702]: E1125 14:11:59.849563 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdf8d623-f3c0-4dc1-a034-f9a10c44ac39" containerName="glance-log" Nov 25 14:11:59 crc kubenswrapper[4702]: I1125 14:11:59.849641 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdf8d623-f3c0-4dc1-a034-f9a10c44ac39" containerName="glance-log" Nov 25 14:11:59 crc kubenswrapper[4702]: E1125 14:11:59.849805 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdf8d623-f3c0-4dc1-a034-f9a10c44ac39" containerName="glance-httpd" Nov 25 14:11:59 crc kubenswrapper[4702]: I1125 14:11:59.849901 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdf8d623-f3c0-4dc1-a034-f9a10c44ac39" containerName="glance-httpd" Nov 25 14:11:59 crc kubenswrapper[4702]: E1125 14:11:59.849997 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cecda5cc-95e9-4966-a1dd-702cb78e4623" containerName="registry-server" Nov 25 14:11:59 crc kubenswrapper[4702]: I1125 14:11:59.850073 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="cecda5cc-95e9-4966-a1dd-702cb78e4623" containerName="registry-server" Nov 25 14:11:59 crc kubenswrapper[4702]: E1125 14:11:59.850164 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc2d6b57-c30a-4d66-a0d1-0cdc31e24309" containerName="glance-httpd" Nov 25 14:11:59 crc kubenswrapper[4702]: I1125 14:11:59.850314 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc2d6b57-c30a-4d66-a0d1-0cdc31e24309" containerName="glance-httpd" Nov 25 14:11:59 crc kubenswrapper[4702]: E1125 14:11:59.850419 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cecda5cc-95e9-4966-a1dd-702cb78e4623" containerName="extract-content" Nov 25 14:11:59 crc kubenswrapper[4702]: I1125 14:11:59.850497 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="cecda5cc-95e9-4966-a1dd-702cb78e4623" containerName="extract-content" Nov 25 14:11:59 crc kubenswrapper[4702]: E1125 14:11:59.850583 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc2d6b57-c30a-4d66-a0d1-0cdc31e24309" containerName="glance-api" Nov 25 14:11:59 crc kubenswrapper[4702]: I1125 14:11:59.850684 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc2d6b57-c30a-4d66-a0d1-0cdc31e24309" containerName="glance-api" Nov 25 14:11:59 crc kubenswrapper[4702]: E1125 14:11:59.850778 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc2d6b57-c30a-4d66-a0d1-0cdc31e24309" containerName="glance-log" Nov 25 14:11:59 crc kubenswrapper[4702]: I1125 14:11:59.850887 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc2d6b57-c30a-4d66-a0d1-0cdc31e24309" containerName="glance-log" Nov 25 14:11:59 crc kubenswrapper[4702]: I1125 14:11:59.851301 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdf8d623-f3c0-4dc1-a034-f9a10c44ac39" containerName="glance-httpd" Nov 25 14:11:59 crc kubenswrapper[4702]: I1125 14:11:59.851410 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdf8d623-f3c0-4dc1-a034-f9a10c44ac39" containerName="glance-api" Nov 25 14:11:59 crc kubenswrapper[4702]: I1125 14:11:59.851495 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdf8d623-f3c0-4dc1-a034-f9a10c44ac39" containerName="glance-log" Nov 25 14:11:59 crc kubenswrapper[4702]: I1125 14:11:59.851575 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="cecda5cc-95e9-4966-a1dd-702cb78e4623" containerName="registry-server" Nov 25 14:11:59 crc kubenswrapper[4702]: I1125 14:11:59.851654 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc2d6b57-c30a-4d66-a0d1-0cdc31e24309" containerName="glance-api" Nov 25 14:11:59 crc kubenswrapper[4702]: I1125 14:11:59.851747 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc2d6b57-c30a-4d66-a0d1-0cdc31e24309" containerName="glance-log" Nov 25 14:11:59 crc kubenswrapper[4702]: I1125 14:11:59.851835 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc2d6b57-c30a-4d66-a0d1-0cdc31e24309" containerName="glance-httpd" Nov 25 14:11:59 crc kubenswrapper[4702]: I1125 14:11:59.854093 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2sz5r" Nov 25 14:11:59 crc kubenswrapper[4702]: I1125 14:11:59.876050 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2sz5r"] Nov 25 14:12:00 crc kubenswrapper[4702]: I1125 14:12:00.012677 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/018f6a3f-6536-4b89-84ee-c3b40c83ec64-utilities\") pod \"redhat-marketplace-2sz5r\" (UID: \"018f6a3f-6536-4b89-84ee-c3b40c83ec64\") " pod="openshift-marketplace/redhat-marketplace-2sz5r" Nov 25 14:12:00 crc kubenswrapper[4702]: I1125 14:12:00.012740 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89prl\" (UniqueName: \"kubernetes.io/projected/018f6a3f-6536-4b89-84ee-c3b40c83ec64-kube-api-access-89prl\") pod \"redhat-marketplace-2sz5r\" (UID: \"018f6a3f-6536-4b89-84ee-c3b40c83ec64\") " pod="openshift-marketplace/redhat-marketplace-2sz5r" Nov 25 14:12:00 crc kubenswrapper[4702]: I1125 14:12:00.012768 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/018f6a3f-6536-4b89-84ee-c3b40c83ec64-catalog-content\") pod \"redhat-marketplace-2sz5r\" (UID: \"018f6a3f-6536-4b89-84ee-c3b40c83ec64\") " pod="openshift-marketplace/redhat-marketplace-2sz5r" Nov 25 14:12:00 crc kubenswrapper[4702]: I1125 14:12:00.115652 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/018f6a3f-6536-4b89-84ee-c3b40c83ec64-utilities\") pod \"redhat-marketplace-2sz5r\" (UID: \"018f6a3f-6536-4b89-84ee-c3b40c83ec64\") " pod="openshift-marketplace/redhat-marketplace-2sz5r" Nov 25 14:12:00 crc kubenswrapper[4702]: I1125 14:12:00.115719 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89prl\" (UniqueName: \"kubernetes.io/projected/018f6a3f-6536-4b89-84ee-c3b40c83ec64-kube-api-access-89prl\") pod \"redhat-marketplace-2sz5r\" (UID: \"018f6a3f-6536-4b89-84ee-c3b40c83ec64\") " pod="openshift-marketplace/redhat-marketplace-2sz5r" Nov 25 14:12:00 crc kubenswrapper[4702]: I1125 14:12:00.115749 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/018f6a3f-6536-4b89-84ee-c3b40c83ec64-catalog-content\") pod \"redhat-marketplace-2sz5r\" (UID: \"018f6a3f-6536-4b89-84ee-c3b40c83ec64\") " pod="openshift-marketplace/redhat-marketplace-2sz5r" Nov 25 14:12:00 crc kubenswrapper[4702]: I1125 14:12:00.116512 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/018f6a3f-6536-4b89-84ee-c3b40c83ec64-catalog-content\") pod \"redhat-marketplace-2sz5r\" (UID: \"018f6a3f-6536-4b89-84ee-c3b40c83ec64\") " pod="openshift-marketplace/redhat-marketplace-2sz5r" Nov 25 14:12:00 crc kubenswrapper[4702]: I1125 14:12:00.116873 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/018f6a3f-6536-4b89-84ee-c3b40c83ec64-utilities\") pod \"redhat-marketplace-2sz5r\" (UID: \"018f6a3f-6536-4b89-84ee-c3b40c83ec64\") " pod="openshift-marketplace/redhat-marketplace-2sz5r" Nov 25 14:12:00 crc kubenswrapper[4702]: I1125 14:12:00.141212 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89prl\" (UniqueName: \"kubernetes.io/projected/018f6a3f-6536-4b89-84ee-c3b40c83ec64-kube-api-access-89prl\") pod \"redhat-marketplace-2sz5r\" (UID: \"018f6a3f-6536-4b89-84ee-c3b40c83ec64\") " pod="openshift-marketplace/redhat-marketplace-2sz5r" Nov 25 14:12:00 crc kubenswrapper[4702]: I1125 14:12:00.177970 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2sz5r" Nov 25 14:12:00 crc kubenswrapper[4702]: I1125 14:12:00.470119 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2sz5r"] Nov 25 14:12:01 crc kubenswrapper[4702]: I1125 14:12:01.039757 4702 generic.go:334] "Generic (PLEG): container finished" podID="018f6a3f-6536-4b89-84ee-c3b40c83ec64" containerID="d6685504131f6d069b076af2f04beecc88ca72976c0c8526213057a434e27228" exitCode=0 Nov 25 14:12:01 crc kubenswrapper[4702]: I1125 14:12:01.039848 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2sz5r" event={"ID":"018f6a3f-6536-4b89-84ee-c3b40c83ec64","Type":"ContainerDied","Data":"d6685504131f6d069b076af2f04beecc88ca72976c0c8526213057a434e27228"} Nov 25 14:12:01 crc kubenswrapper[4702]: I1125 14:12:01.039947 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2sz5r" event={"ID":"018f6a3f-6536-4b89-84ee-c3b40c83ec64","Type":"ContainerStarted","Data":"206e67a49cbcdf724ff5f213425afc649a94ca3a39608b7e54638ed00b32aeed"} Nov 25 14:12:03 crc kubenswrapper[4702]: I1125 14:12:03.066745 4702 generic.go:334] "Generic (PLEG): container finished" podID="018f6a3f-6536-4b89-84ee-c3b40c83ec64" containerID="5b9dee8950497464efa4f0faafdf99b3039c4b04e61af031b773b317d0b6992b" exitCode=0 Nov 25 14:12:03 crc kubenswrapper[4702]: I1125 14:12:03.066811 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2sz5r" event={"ID":"018f6a3f-6536-4b89-84ee-c3b40c83ec64","Type":"ContainerDied","Data":"5b9dee8950497464efa4f0faafdf99b3039c4b04e61af031b773b317d0b6992b"} Nov 25 14:12:05 crc kubenswrapper[4702]: E1125 14:12:05.014405 4702 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcecda5cc_95e9_4966_a1dd_702cb78e4623.slice/crio-c6a84396dbcaa8e9796180d0d842af7b05bae72486b7c07eae992e3ac3f78bb5.scope\": RecentStats: unable to find data in memory cache]" Nov 25 14:12:05 crc kubenswrapper[4702]: I1125 14:12:05.085267 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2sz5r" event={"ID":"018f6a3f-6536-4b89-84ee-c3b40c83ec64","Type":"ContainerStarted","Data":"c6f09f1ec72ff65ae862e9f0c44966eb47191abaf8cc7c1b6eb80df3c58b24ef"} Nov 25 14:12:05 crc kubenswrapper[4702]: I1125 14:12:05.119054 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2sz5r" podStartSLOduration=3.047221299 podStartE2EDuration="6.119015426s" podCreationTimestamp="2025-11-25 14:11:59 +0000 UTC" firstStartedPulling="2025-11-25 14:12:01.044124227 +0000 UTC m=+1530.206774677" lastFinishedPulling="2025-11-25 14:12:04.115918354 +0000 UTC m=+1533.278568804" observedRunningTime="2025-11-25 14:12:05.111496189 +0000 UTC m=+1534.274146649" watchObservedRunningTime="2025-11-25 14:12:05.119015426 +0000 UTC m=+1534.281665876" Nov 25 14:12:10 crc kubenswrapper[4702]: I1125 14:12:10.178918 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2sz5r" Nov 25 14:12:10 crc kubenswrapper[4702]: I1125 14:12:10.180035 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2sz5r" Nov 25 14:12:10 crc kubenswrapper[4702]: I1125 14:12:10.248684 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2sz5r" Nov 25 14:12:11 crc kubenswrapper[4702]: I1125 14:12:11.230558 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2sz5r" Nov 25 14:12:11 crc kubenswrapper[4702]: I1125 14:12:11.289964 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2sz5r"] Nov 25 14:12:13 crc kubenswrapper[4702]: I1125 14:12:13.203038 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2sz5r" podUID="018f6a3f-6536-4b89-84ee-c3b40c83ec64" containerName="registry-server" containerID="cri-o://c6f09f1ec72ff65ae862e9f0c44966eb47191abaf8cc7c1b6eb80df3c58b24ef" gracePeriod=2 Nov 25 14:12:13 crc kubenswrapper[4702]: I1125 14:12:13.664479 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2sz5r" Nov 25 14:12:13 crc kubenswrapper[4702]: I1125 14:12:13.732376 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/018f6a3f-6536-4b89-84ee-c3b40c83ec64-catalog-content\") pod \"018f6a3f-6536-4b89-84ee-c3b40c83ec64\" (UID: \"018f6a3f-6536-4b89-84ee-c3b40c83ec64\") " Nov 25 14:12:13 crc kubenswrapper[4702]: I1125 14:12:13.732498 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89prl\" (UniqueName: \"kubernetes.io/projected/018f6a3f-6536-4b89-84ee-c3b40c83ec64-kube-api-access-89prl\") pod \"018f6a3f-6536-4b89-84ee-c3b40c83ec64\" (UID: \"018f6a3f-6536-4b89-84ee-c3b40c83ec64\") " Nov 25 14:12:13 crc kubenswrapper[4702]: I1125 14:12:13.732595 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/018f6a3f-6536-4b89-84ee-c3b40c83ec64-utilities\") pod \"018f6a3f-6536-4b89-84ee-c3b40c83ec64\" (UID: \"018f6a3f-6536-4b89-84ee-c3b40c83ec64\") " Nov 25 14:12:13 crc kubenswrapper[4702]: I1125 14:12:13.733762 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/018f6a3f-6536-4b89-84ee-c3b40c83ec64-utilities" (OuterVolumeSpecName: "utilities") pod "018f6a3f-6536-4b89-84ee-c3b40c83ec64" (UID: "018f6a3f-6536-4b89-84ee-c3b40c83ec64"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:12:13 crc kubenswrapper[4702]: I1125 14:12:13.740830 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/018f6a3f-6536-4b89-84ee-c3b40c83ec64-kube-api-access-89prl" (OuterVolumeSpecName: "kube-api-access-89prl") pod "018f6a3f-6536-4b89-84ee-c3b40c83ec64" (UID: "018f6a3f-6536-4b89-84ee-c3b40c83ec64"). InnerVolumeSpecName "kube-api-access-89prl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:12:13 crc kubenswrapper[4702]: I1125 14:12:13.758345 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/018f6a3f-6536-4b89-84ee-c3b40c83ec64-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "018f6a3f-6536-4b89-84ee-c3b40c83ec64" (UID: "018f6a3f-6536-4b89-84ee-c3b40c83ec64"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:12:13 crc kubenswrapper[4702]: I1125 14:12:13.836065 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/018f6a3f-6536-4b89-84ee-c3b40c83ec64-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 14:12:13 crc kubenswrapper[4702]: I1125 14:12:13.836110 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89prl\" (UniqueName: \"kubernetes.io/projected/018f6a3f-6536-4b89-84ee-c3b40c83ec64-kube-api-access-89prl\") on node \"crc\" DevicePath \"\"" Nov 25 14:12:13 crc kubenswrapper[4702]: I1125 14:12:13.836124 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/018f6a3f-6536-4b89-84ee-c3b40c83ec64-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 14:12:14 crc kubenswrapper[4702]: I1125 14:12:14.215629 4702 generic.go:334] "Generic (PLEG): container finished" podID="018f6a3f-6536-4b89-84ee-c3b40c83ec64" containerID="c6f09f1ec72ff65ae862e9f0c44966eb47191abaf8cc7c1b6eb80df3c58b24ef" exitCode=0 Nov 25 14:12:14 crc kubenswrapper[4702]: I1125 14:12:14.215744 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2sz5r" Nov 25 14:12:14 crc kubenswrapper[4702]: I1125 14:12:14.215752 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2sz5r" event={"ID":"018f6a3f-6536-4b89-84ee-c3b40c83ec64","Type":"ContainerDied","Data":"c6f09f1ec72ff65ae862e9f0c44966eb47191abaf8cc7c1b6eb80df3c58b24ef"} Nov 25 14:12:14 crc kubenswrapper[4702]: I1125 14:12:14.216467 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2sz5r" event={"ID":"018f6a3f-6536-4b89-84ee-c3b40c83ec64","Type":"ContainerDied","Data":"206e67a49cbcdf724ff5f213425afc649a94ca3a39608b7e54638ed00b32aeed"} Nov 25 14:12:14 crc kubenswrapper[4702]: I1125 14:12:14.216514 4702 scope.go:117] "RemoveContainer" containerID="c6f09f1ec72ff65ae862e9f0c44966eb47191abaf8cc7c1b6eb80df3c58b24ef" Nov 25 14:12:14 crc kubenswrapper[4702]: I1125 14:12:14.247296 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2sz5r"] Nov 25 14:12:14 crc kubenswrapper[4702]: I1125 14:12:14.252986 4702 scope.go:117] "RemoveContainer" containerID="5b9dee8950497464efa4f0faafdf99b3039c4b04e61af031b773b317d0b6992b" Nov 25 14:12:14 crc kubenswrapper[4702]: I1125 14:12:14.254696 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2sz5r"] Nov 25 14:12:14 crc kubenswrapper[4702]: I1125 14:12:14.280364 4702 scope.go:117] "RemoveContainer" containerID="d6685504131f6d069b076af2f04beecc88ca72976c0c8526213057a434e27228" Nov 25 14:12:14 crc kubenswrapper[4702]: I1125 14:12:14.302465 4702 scope.go:117] "RemoveContainer" containerID="c6f09f1ec72ff65ae862e9f0c44966eb47191abaf8cc7c1b6eb80df3c58b24ef" Nov 25 14:12:14 crc kubenswrapper[4702]: E1125 14:12:14.302999 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6f09f1ec72ff65ae862e9f0c44966eb47191abaf8cc7c1b6eb80df3c58b24ef\": container with ID starting with c6f09f1ec72ff65ae862e9f0c44966eb47191abaf8cc7c1b6eb80df3c58b24ef not found: ID does not exist" containerID="c6f09f1ec72ff65ae862e9f0c44966eb47191abaf8cc7c1b6eb80df3c58b24ef" Nov 25 14:12:14 crc kubenswrapper[4702]: I1125 14:12:14.303057 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6f09f1ec72ff65ae862e9f0c44966eb47191abaf8cc7c1b6eb80df3c58b24ef"} err="failed to get container status \"c6f09f1ec72ff65ae862e9f0c44966eb47191abaf8cc7c1b6eb80df3c58b24ef\": rpc error: code = NotFound desc = could not find container \"c6f09f1ec72ff65ae862e9f0c44966eb47191abaf8cc7c1b6eb80df3c58b24ef\": container with ID starting with c6f09f1ec72ff65ae862e9f0c44966eb47191abaf8cc7c1b6eb80df3c58b24ef not found: ID does not exist" Nov 25 14:12:14 crc kubenswrapper[4702]: I1125 14:12:14.303097 4702 scope.go:117] "RemoveContainer" containerID="5b9dee8950497464efa4f0faafdf99b3039c4b04e61af031b773b317d0b6992b" Nov 25 14:12:14 crc kubenswrapper[4702]: E1125 14:12:14.303421 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b9dee8950497464efa4f0faafdf99b3039c4b04e61af031b773b317d0b6992b\": container with ID starting with 5b9dee8950497464efa4f0faafdf99b3039c4b04e61af031b773b317d0b6992b not found: ID does not exist" containerID="5b9dee8950497464efa4f0faafdf99b3039c4b04e61af031b773b317d0b6992b" Nov 25 14:12:14 crc kubenswrapper[4702]: I1125 14:12:14.303459 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b9dee8950497464efa4f0faafdf99b3039c4b04e61af031b773b317d0b6992b"} err="failed to get container status \"5b9dee8950497464efa4f0faafdf99b3039c4b04e61af031b773b317d0b6992b\": rpc error: code = NotFound desc = could not find container \"5b9dee8950497464efa4f0faafdf99b3039c4b04e61af031b773b317d0b6992b\": container with ID starting with 5b9dee8950497464efa4f0faafdf99b3039c4b04e61af031b773b317d0b6992b not found: ID does not exist" Nov 25 14:12:14 crc kubenswrapper[4702]: I1125 14:12:14.303478 4702 scope.go:117] "RemoveContainer" containerID="d6685504131f6d069b076af2f04beecc88ca72976c0c8526213057a434e27228" Nov 25 14:12:14 crc kubenswrapper[4702]: E1125 14:12:14.303745 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6685504131f6d069b076af2f04beecc88ca72976c0c8526213057a434e27228\": container with ID starting with d6685504131f6d069b076af2f04beecc88ca72976c0c8526213057a434e27228 not found: ID does not exist" containerID="d6685504131f6d069b076af2f04beecc88ca72976c0c8526213057a434e27228" Nov 25 14:12:14 crc kubenswrapper[4702]: I1125 14:12:14.303776 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6685504131f6d069b076af2f04beecc88ca72976c0c8526213057a434e27228"} err="failed to get container status \"d6685504131f6d069b076af2f04beecc88ca72976c0c8526213057a434e27228\": rpc error: code = NotFound desc = could not find container \"d6685504131f6d069b076af2f04beecc88ca72976c0c8526213057a434e27228\": container with ID starting with d6685504131f6d069b076af2f04beecc88ca72976c0c8526213057a434e27228 not found: ID does not exist" Nov 25 14:12:15 crc kubenswrapper[4702]: E1125 14:12:15.242546 4702 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcecda5cc_95e9_4966_a1dd_702cb78e4623.slice/crio-c6a84396dbcaa8e9796180d0d842af7b05bae72486b7c07eae992e3ac3f78bb5.scope\": RecentStats: unable to find data in memory cache]" Nov 25 14:12:15 crc kubenswrapper[4702]: I1125 14:12:15.818403 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="018f6a3f-6536-4b89-84ee-c3b40c83ec64" path="/var/lib/kubelet/pods/018f6a3f-6536-4b89-84ee-c3b40c83ec64/volumes" Nov 25 14:12:25 crc kubenswrapper[4702]: E1125 14:12:25.428902 4702 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcecda5cc_95e9_4966_a1dd_702cb78e4623.slice/crio-c6a84396dbcaa8e9796180d0d842af7b05bae72486b7c07eae992e3ac3f78bb5.scope\": RecentStats: unable to find data in memory cache]" Nov 25 14:12:31 crc kubenswrapper[4702]: E1125 14:12:31.811844 4702 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/380a904223ea3a1ceada69ec7c3030ebb94e4ff08d034fc42c5f95798184f614/diff" to get inode usage: stat /var/lib/containers/storage/overlay/380a904223ea3a1ceada69ec7c3030ebb94e4ff08d034fc42c5f95798184f614/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openshift-marketplace_community-operators-dqgk4_cecda5cc-95e9-4966-a1dd-702cb78e4623/registry-server/0.log" to get inode usage: stat /var/log/pods/openshift-marketplace_community-operators-dqgk4_cecda5cc-95e9-4966-a1dd-702cb78e4623/registry-server/0.log: no such file or directory Nov 25 14:12:48 crc kubenswrapper[4702]: I1125 14:12:48.070354 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-create-smlsw"] Nov 25 14:12:48 crc kubenswrapper[4702]: I1125 14:12:48.080924 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-1d57-account-create-update-rnh45"] Nov 25 14:12:48 crc kubenswrapper[4702]: I1125 14:12:48.090167 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-create-smlsw"] Nov 25 14:12:48 crc kubenswrapper[4702]: I1125 14:12:48.098743 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-1d57-account-create-update-rnh45"] Nov 25 14:12:49 crc kubenswrapper[4702]: I1125 14:12:49.811588 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d45b3f0-68f4-488f-b356-660f02a504ee" path="/var/lib/kubelet/pods/8d45b3f0-68f4-488f-b356-660f02a504ee/volumes" Nov 25 14:12:49 crc kubenswrapper[4702]: I1125 14:12:49.813174 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a144271-8855-4fba-8913-d9e1cef952d0" path="/var/lib/kubelet/pods/9a144271-8855-4fba-8913-d9e1cef952d0/volumes" Nov 25 14:13:26 crc kubenswrapper[4702]: I1125 14:13:26.061318 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-5j9zq"] Nov 25 14:13:26 crc kubenswrapper[4702]: I1125 14:13:26.069937 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-5j9zq"] Nov 25 14:13:27 crc kubenswrapper[4702]: I1125 14:13:27.816591 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3339fdec-c17b-4052-a472-7c6e19475fe4" path="/var/lib/kubelet/pods/3339fdec-c17b-4052-a472-7c6e19475fe4/volumes" Nov 25 14:13:33 crc kubenswrapper[4702]: I1125 14:13:33.042200 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-88sbx"] Nov 25 14:13:33 crc kubenswrapper[4702]: I1125 14:13:33.052476 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-88sbx"] Nov 25 14:13:33 crc kubenswrapper[4702]: I1125 14:13:33.228081 4702 scope.go:117] "RemoveContainer" containerID="a7bc21f37028e958666ea2c6c8d71a826c4871f9071098876c4467c088b82dc0" Nov 25 14:13:33 crc kubenswrapper[4702]: I1125 14:13:33.252452 4702 scope.go:117] "RemoveContainer" containerID="f336a0efc3b1011c580f7efb9b0362e23c98ae426a5226b3ecb6e71a1275a841" Nov 25 14:13:33 crc kubenswrapper[4702]: I1125 14:13:33.278039 4702 scope.go:117] "RemoveContainer" containerID="022275568c8772696cf7fba610a0bcdb3ce8f631d918ee164271905aa1bf0b10" Nov 25 14:13:33 crc kubenswrapper[4702]: I1125 14:13:33.308862 4702 scope.go:117] "RemoveContainer" containerID="c693c08b947e18d663547be50ad70954bc994cf9faf3218e24e1713f65ea23a2" Nov 25 14:13:33 crc kubenswrapper[4702]: I1125 14:13:33.339522 4702 scope.go:117] "RemoveContainer" containerID="2af2837b46377b1533459d4a0613c8e64b3a32f54338f03ae72128307b90e344" Nov 25 14:13:33 crc kubenswrapper[4702]: I1125 14:13:33.393610 4702 scope.go:117] "RemoveContainer" containerID="e319cefabb880657e7e32ed663cba7b7067cd13a0d028da5e0f17f9faa7857f9" Nov 25 14:13:33 crc kubenswrapper[4702]: I1125 14:13:33.413096 4702 scope.go:117] "RemoveContainer" containerID="ab70360378ad003c37db0e35718f2a0c9d8610dab59da2005e4185453011362e" Nov 25 14:13:33 crc kubenswrapper[4702]: I1125 14:13:33.810504 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12d991ff-0e51-4ed3-8675-b3c1123bbda0" path="/var/lib/kubelet/pods/12d991ff-0e51-4ed3-8675-b3c1123bbda0/volumes" Nov 25 14:13:33 crc kubenswrapper[4702]: I1125 14:13:33.962898 4702 patch_prober.go:28] interesting pod/machine-config-daemon-qqr76 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 14:13:33 crc kubenswrapper[4702]: I1125 14:13:33.962996 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 14:14:03 crc kubenswrapper[4702]: I1125 14:14:03.963597 4702 patch_prober.go:28] interesting pod/machine-config-daemon-qqr76 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 14:14:03 crc kubenswrapper[4702]: I1125 14:14:03.964635 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.401500 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 14:14:25 crc kubenswrapper[4702]: E1125 14:14:25.402506 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="018f6a3f-6536-4b89-84ee-c3b40c83ec64" containerName="extract-utilities" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.402523 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="018f6a3f-6536-4b89-84ee-c3b40c83ec64" containerName="extract-utilities" Nov 25 14:14:25 crc kubenswrapper[4702]: E1125 14:14:25.402545 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="018f6a3f-6536-4b89-84ee-c3b40c83ec64" containerName="extract-content" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.402551 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="018f6a3f-6536-4b89-84ee-c3b40c83ec64" containerName="extract-content" Nov 25 14:14:25 crc kubenswrapper[4702]: E1125 14:14:25.402569 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="018f6a3f-6536-4b89-84ee-c3b40c83ec64" containerName="registry-server" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.402576 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="018f6a3f-6536-4b89-84ee-c3b40c83ec64" containerName="registry-server" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.402722 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="018f6a3f-6536-4b89-84ee-c3b40c83ec64" containerName="registry-server" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.403799 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.434218 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.520065 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-dev\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.520160 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-sys\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.520395 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.520449 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-run\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.520506 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.520545 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.520656 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.520684 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.520721 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b5b9839-7217-4d14-a6e3-c8e62579f40e-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.520751 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b5b9839-7217-4d14-a6e3-c8e62579f40e-scripts\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.520861 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.520907 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b5b9839-7217-4d14-a6e3-c8e62579f40e-logs\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.520975 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lfn2\" (UniqueName: \"kubernetes.io/projected/2b5b9839-7217-4d14-a6e3-c8e62579f40e-kube-api-access-4lfn2\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.521107 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b5b9839-7217-4d14-a6e3-c8e62579f40e-config-data\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.527873 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.529102 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.531537 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"openstack-config-secret" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.532192 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts-9db6gc427h" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.533414 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.533471 4702 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"default-dockercfg-cc8rt" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.541127 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.623026 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.623094 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b5b9839-7217-4d14-a6e3-c8e62579f40e-logs\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.623143 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lfn2\" (UniqueName: \"kubernetes.io/projected/2b5b9839-7217-4d14-a6e3-c8e62579f40e-kube-api-access-4lfn2\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.623182 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b5b9839-7217-4d14-a6e3-c8e62579f40e-config-data\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.623205 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-dev\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.623249 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9366d936-2d43-4a0b-a8e6-4006afebd985-openstack-config\") pod \"openstackclient\" (UID: \"9366d936-2d43-4a0b-a8e6-4006afebd985\") " pod="glance-kuttl-tests/openstackclient" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.623272 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-sys\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.623299 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-run\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.623319 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.623340 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.623359 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.623389 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9366d936-2d43-4a0b-a8e6-4006afebd985-openstack-config-secret\") pod \"openstackclient\" (UID: \"9366d936-2d43-4a0b-a8e6-4006afebd985\") " pod="glance-kuttl-tests/openstackclient" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.623409 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.623425 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.623441 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b5b9839-7217-4d14-a6e3-c8e62579f40e-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.623458 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b5b9839-7217-4d14-a6e3-c8e62579f40e-scripts\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.623476 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phtrd\" (UniqueName: \"kubernetes.io/projected/9366d936-2d43-4a0b-a8e6-4006afebd985-kube-api-access-phtrd\") pod \"openstackclient\" (UID: \"9366d936-2d43-4a0b-a8e6-4006afebd985\") " pod="glance-kuttl-tests/openstackclient" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.623494 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/9366d936-2d43-4a0b-a8e6-4006afebd985-openstack-scripts\") pod \"openstackclient\" (UID: \"9366d936-2d43-4a0b-a8e6-4006afebd985\") " pod="glance-kuttl-tests/openstackclient" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.623638 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.623960 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.624045 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.624047 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.624060 4702 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.624093 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.624109 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-dev\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.624126 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-sys\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.624311 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-run\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.624509 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b5b9839-7217-4d14-a6e3-c8e62579f40e-logs\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.626329 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b5b9839-7217-4d14-a6e3-c8e62579f40e-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.634169 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b5b9839-7217-4d14-a6e3-c8e62579f40e-scripts\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.636804 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b5b9839-7217-4d14-a6e3-c8e62579f40e-config-data\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.659734 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lfn2\" (UniqueName: \"kubernetes.io/projected/2b5b9839-7217-4d14-a6e3-c8e62579f40e-kube-api-access-4lfn2\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.663855 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.669324 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-1\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.721811 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.725019 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9366d936-2d43-4a0b-a8e6-4006afebd985-openstack-config\") pod \"openstackclient\" (UID: \"9366d936-2d43-4a0b-a8e6-4006afebd985\") " pod="glance-kuttl-tests/openstackclient" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.725079 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9366d936-2d43-4a0b-a8e6-4006afebd985-openstack-config-secret\") pod \"openstackclient\" (UID: \"9366d936-2d43-4a0b-a8e6-4006afebd985\") " pod="glance-kuttl-tests/openstackclient" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.725102 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phtrd\" (UniqueName: \"kubernetes.io/projected/9366d936-2d43-4a0b-a8e6-4006afebd985-kube-api-access-phtrd\") pod \"openstackclient\" (UID: \"9366d936-2d43-4a0b-a8e6-4006afebd985\") " pod="glance-kuttl-tests/openstackclient" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.725131 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/9366d936-2d43-4a0b-a8e6-4006afebd985-openstack-scripts\") pod \"openstackclient\" (UID: \"9366d936-2d43-4a0b-a8e6-4006afebd985\") " pod="glance-kuttl-tests/openstackclient" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.726062 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/9366d936-2d43-4a0b-a8e6-4006afebd985-openstack-scripts\") pod \"openstackclient\" (UID: \"9366d936-2d43-4a0b-a8e6-4006afebd985\") " pod="glance-kuttl-tests/openstackclient" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.726564 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9366d936-2d43-4a0b-a8e6-4006afebd985-openstack-config\") pod \"openstackclient\" (UID: \"9366d936-2d43-4a0b-a8e6-4006afebd985\") " pod="glance-kuttl-tests/openstackclient" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.730423 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9366d936-2d43-4a0b-a8e6-4006afebd985-openstack-config-secret\") pod \"openstackclient\" (UID: \"9366d936-2d43-4a0b-a8e6-4006afebd985\") " pod="glance-kuttl-tests/openstackclient" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.745485 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phtrd\" (UniqueName: \"kubernetes.io/projected/9366d936-2d43-4a0b-a8e6-4006afebd985-kube-api-access-phtrd\") pod \"openstackclient\" (UID: \"9366d936-2d43-4a0b-a8e6-4006afebd985\") " pod="glance-kuttl-tests/openstackclient" Nov 25 14:14:25 crc kubenswrapper[4702]: I1125 14:14:25.857667 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 25 14:14:26 crc kubenswrapper[4702]: I1125 14:14:26.191265 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 14:14:26 crc kubenswrapper[4702]: I1125 14:14:26.305617 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 14:14:26 crc kubenswrapper[4702]: W1125 14:14:26.313986 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9366d936_2d43_4a0b_a8e6_4006afebd985.slice/crio-391b861e2016e05f23f8670d84ca4aff47860a18702f42f143f49cc74af9249e WatchSource:0}: Error finding container 391b861e2016e05f23f8670d84ca4aff47860a18702f42f143f49cc74af9249e: Status 404 returned error can't find the container with id 391b861e2016e05f23f8670d84ca4aff47860a18702f42f143f49cc74af9249e Nov 25 14:14:26 crc kubenswrapper[4702]: I1125 14:14:26.433978 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"9366d936-2d43-4a0b-a8e6-4006afebd985","Type":"ContainerStarted","Data":"391b861e2016e05f23f8670d84ca4aff47860a18702f42f143f49cc74af9249e"} Nov 25 14:14:26 crc kubenswrapper[4702]: I1125 14:14:26.436487 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"2b5b9839-7217-4d14-a6e3-c8e62579f40e","Type":"ContainerStarted","Data":"bae79b726b36329b7fbf6dae1150e950506789af04b0c406f4327ea2b7bf7732"} Nov 25 14:14:27 crc kubenswrapper[4702]: I1125 14:14:27.446849 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"9366d936-2d43-4a0b-a8e6-4006afebd985","Type":"ContainerStarted","Data":"68d05adc30aa651355c13174f8c3106716a75de738f81febe7a6ac6ebaa18f73"} Nov 25 14:14:27 crc kubenswrapper[4702]: I1125 14:14:27.449988 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"2b5b9839-7217-4d14-a6e3-c8e62579f40e","Type":"ContainerStarted","Data":"c91838e0db075b4ce25a4f519b9aa59e579ebf0f2e0c6fd0aadaf23dae87e9c6"} Nov 25 14:14:27 crc kubenswrapper[4702]: I1125 14:14:27.450040 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"2b5b9839-7217-4d14-a6e3-c8e62579f40e","Type":"ContainerStarted","Data":"3e26d313616b3a70fe756a3d41215ec16acd28c473c3365fe61988d4e73729a1"} Nov 25 14:14:27 crc kubenswrapper[4702]: I1125 14:14:27.450054 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"2b5b9839-7217-4d14-a6e3-c8e62579f40e","Type":"ContainerStarted","Data":"178cd9604dee844af6359851c1ff829c6dc3b24c42e2b65fd64e216a696ca23d"} Nov 25 14:14:27 crc kubenswrapper[4702]: I1125 14:14:27.480980 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstackclient" podStartSLOduration=2.480942105 podStartE2EDuration="2.480942105s" podCreationTimestamp="2025-11-25 14:14:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 14:14:27.467711237 +0000 UTC m=+1676.630361697" watchObservedRunningTime="2025-11-25 14:14:27.480942105 +0000 UTC m=+1676.643592575" Nov 25 14:14:27 crc kubenswrapper[4702]: I1125 14:14:27.511582 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-1" podStartSLOduration=2.511553375 podStartE2EDuration="2.511553375s" podCreationTimestamp="2025-11-25 14:14:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 14:14:27.50166119 +0000 UTC m=+1676.664311640" watchObservedRunningTime="2025-11-25 14:14:27.511553375 +0000 UTC m=+1676.674203825" Nov 25 14:14:33 crc kubenswrapper[4702]: I1125 14:14:33.580736 4702 scope.go:117] "RemoveContainer" containerID="3112282e6ea4f2f10dd9003521c9711e8ee572d8eb0a9ae8a3e98e9aa34b5888" Nov 25 14:14:33 crc kubenswrapper[4702]: I1125 14:14:33.615460 4702 scope.go:117] "RemoveContainer" containerID="bcfe02a179b6890c70d2d8c7f50bc272681575257f235aee7da8e4e3d0c57fe6" Nov 25 14:14:33 crc kubenswrapper[4702]: I1125 14:14:33.963177 4702 patch_prober.go:28] interesting pod/machine-config-daemon-qqr76 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 14:14:33 crc kubenswrapper[4702]: I1125 14:14:33.963278 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 14:14:33 crc kubenswrapper[4702]: I1125 14:14:33.963322 4702 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" Nov 25 14:14:33 crc kubenswrapper[4702]: I1125 14:14:33.963936 4702 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"32273717263d7fcc99e8193a439dd976da66869c55bb4633d9314cacb6ab92a1"} pod="openshift-machine-config-operator/machine-config-daemon-qqr76" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 14:14:33 crc kubenswrapper[4702]: I1125 14:14:33.964007 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" containerID="cri-o://32273717263d7fcc99e8193a439dd976da66869c55bb4633d9314cacb6ab92a1" gracePeriod=600 Nov 25 14:14:34 crc kubenswrapper[4702]: E1125 14:14:34.098543 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qqr76_openshift-machine-config-operator(3c491818-b0bb-4d82-a031-96e2dfac8c27)\"" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" Nov 25 14:14:34 crc kubenswrapper[4702]: I1125 14:14:34.513334 4702 generic.go:334] "Generic (PLEG): container finished" podID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerID="32273717263d7fcc99e8193a439dd976da66869c55bb4633d9314cacb6ab92a1" exitCode=0 Nov 25 14:14:34 crc kubenswrapper[4702]: I1125 14:14:34.513379 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" event={"ID":"3c491818-b0bb-4d82-a031-96e2dfac8c27","Type":"ContainerDied","Data":"32273717263d7fcc99e8193a439dd976da66869c55bb4633d9314cacb6ab92a1"} Nov 25 14:14:34 crc kubenswrapper[4702]: I1125 14:14:34.513408 4702 scope.go:117] "RemoveContainer" containerID="3c4335b16016506091cd0c40c0eeaec31932b0b8e710fd86970dfc2371661ab2" Nov 25 14:14:34 crc kubenswrapper[4702]: I1125 14:14:34.514001 4702 scope.go:117] "RemoveContainer" containerID="32273717263d7fcc99e8193a439dd976da66869c55bb4633d9314cacb6ab92a1" Nov 25 14:14:34 crc kubenswrapper[4702]: E1125 14:14:34.514315 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qqr76_openshift-machine-config-operator(3c491818-b0bb-4d82-a031-96e2dfac8c27)\"" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" Nov 25 14:14:35 crc kubenswrapper[4702]: I1125 14:14:35.722481 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:35 crc kubenswrapper[4702]: I1125 14:14:35.722634 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:35 crc kubenswrapper[4702]: I1125 14:14:35.722658 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:35 crc kubenswrapper[4702]: I1125 14:14:35.747814 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:35 crc kubenswrapper[4702]: I1125 14:14:35.757911 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:35 crc kubenswrapper[4702]: I1125 14:14:35.780195 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:36 crc kubenswrapper[4702]: I1125 14:14:36.535560 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:36 crc kubenswrapper[4702]: I1125 14:14:36.535643 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:36 crc kubenswrapper[4702]: I1125 14:14:36.535657 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:36 crc kubenswrapper[4702]: I1125 14:14:36.549362 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:36 crc kubenswrapper[4702]: I1125 14:14:36.549830 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:36 crc kubenswrapper[4702]: I1125 14:14:36.551370 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:14:45 crc kubenswrapper[4702]: I1125 14:14:45.801443 4702 scope.go:117] "RemoveContainer" containerID="32273717263d7fcc99e8193a439dd976da66869c55bb4633d9314cacb6ab92a1" Nov 25 14:14:45 crc kubenswrapper[4702]: E1125 14:14:45.802181 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qqr76_openshift-machine-config-operator(3c491818-b0bb-4d82-a031-96e2dfac8c27)\"" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" Nov 25 14:14:59 crc kubenswrapper[4702]: I1125 14:14:59.801300 4702 scope.go:117] "RemoveContainer" containerID="32273717263d7fcc99e8193a439dd976da66869c55bb4633d9314cacb6ab92a1" Nov 25 14:14:59 crc kubenswrapper[4702]: E1125 14:14:59.802260 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qqr76_openshift-machine-config-operator(3c491818-b0bb-4d82-a031-96e2dfac8c27)\"" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" Nov 25 14:15:00 crc kubenswrapper[4702]: I1125 14:15:00.145908 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401335-2rddm"] Nov 25 14:15:00 crc kubenswrapper[4702]: I1125 14:15:00.147449 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401335-2rddm" Nov 25 14:15:00 crc kubenswrapper[4702]: I1125 14:15:00.149999 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 14:15:00 crc kubenswrapper[4702]: I1125 14:15:00.149999 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 14:15:00 crc kubenswrapper[4702]: I1125 14:15:00.165373 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401335-2rddm"] Nov 25 14:15:00 crc kubenswrapper[4702]: I1125 14:15:00.237613 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/109f5269-83c5-4ceb-97ab-1eb61a31c4eb-secret-volume\") pod \"collect-profiles-29401335-2rddm\" (UID: \"109f5269-83c5-4ceb-97ab-1eb61a31c4eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401335-2rddm" Nov 25 14:15:00 crc kubenswrapper[4702]: I1125 14:15:00.237763 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npsvm\" (UniqueName: \"kubernetes.io/projected/109f5269-83c5-4ceb-97ab-1eb61a31c4eb-kube-api-access-npsvm\") pod \"collect-profiles-29401335-2rddm\" (UID: \"109f5269-83c5-4ceb-97ab-1eb61a31c4eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401335-2rddm" Nov 25 14:15:00 crc kubenswrapper[4702]: I1125 14:15:00.237839 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/109f5269-83c5-4ceb-97ab-1eb61a31c4eb-config-volume\") pod \"collect-profiles-29401335-2rddm\" (UID: \"109f5269-83c5-4ceb-97ab-1eb61a31c4eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401335-2rddm" Nov 25 14:15:00 crc kubenswrapper[4702]: I1125 14:15:00.339456 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/109f5269-83c5-4ceb-97ab-1eb61a31c4eb-config-volume\") pod \"collect-profiles-29401335-2rddm\" (UID: \"109f5269-83c5-4ceb-97ab-1eb61a31c4eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401335-2rddm" Nov 25 14:15:00 crc kubenswrapper[4702]: I1125 14:15:00.339618 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/109f5269-83c5-4ceb-97ab-1eb61a31c4eb-secret-volume\") pod \"collect-profiles-29401335-2rddm\" (UID: \"109f5269-83c5-4ceb-97ab-1eb61a31c4eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401335-2rddm" Nov 25 14:15:00 crc kubenswrapper[4702]: I1125 14:15:00.339684 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npsvm\" (UniqueName: \"kubernetes.io/projected/109f5269-83c5-4ceb-97ab-1eb61a31c4eb-kube-api-access-npsvm\") pod \"collect-profiles-29401335-2rddm\" (UID: \"109f5269-83c5-4ceb-97ab-1eb61a31c4eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401335-2rddm" Nov 25 14:15:00 crc kubenswrapper[4702]: I1125 14:15:00.340757 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/109f5269-83c5-4ceb-97ab-1eb61a31c4eb-config-volume\") pod \"collect-profiles-29401335-2rddm\" (UID: \"109f5269-83c5-4ceb-97ab-1eb61a31c4eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401335-2rddm" Nov 25 14:15:00 crc kubenswrapper[4702]: I1125 14:15:00.360688 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npsvm\" (UniqueName: \"kubernetes.io/projected/109f5269-83c5-4ceb-97ab-1eb61a31c4eb-kube-api-access-npsvm\") pod \"collect-profiles-29401335-2rddm\" (UID: \"109f5269-83c5-4ceb-97ab-1eb61a31c4eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401335-2rddm" Nov 25 14:15:00 crc kubenswrapper[4702]: I1125 14:15:00.360961 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/109f5269-83c5-4ceb-97ab-1eb61a31c4eb-secret-volume\") pod \"collect-profiles-29401335-2rddm\" (UID: \"109f5269-83c5-4ceb-97ab-1eb61a31c4eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401335-2rddm" Nov 25 14:15:00 crc kubenswrapper[4702]: I1125 14:15:00.472893 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401335-2rddm" Nov 25 14:15:00 crc kubenswrapper[4702]: I1125 14:15:00.888672 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401335-2rddm"] Nov 25 14:15:01 crc kubenswrapper[4702]: I1125 14:15:01.737590 4702 generic.go:334] "Generic (PLEG): container finished" podID="109f5269-83c5-4ceb-97ab-1eb61a31c4eb" containerID="bf418351ab27519d1c1dd2c42c01bce5584b8ee1f5c5bf0fd24b971c31fcef0f" exitCode=0 Nov 25 14:15:01 crc kubenswrapper[4702]: I1125 14:15:01.737650 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401335-2rddm" event={"ID":"109f5269-83c5-4ceb-97ab-1eb61a31c4eb","Type":"ContainerDied","Data":"bf418351ab27519d1c1dd2c42c01bce5584b8ee1f5c5bf0fd24b971c31fcef0f"} Nov 25 14:15:01 crc kubenswrapper[4702]: I1125 14:15:01.737701 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401335-2rddm" event={"ID":"109f5269-83c5-4ceb-97ab-1eb61a31c4eb","Type":"ContainerStarted","Data":"00d99b35a275aea79e4f9ceed4c70f9cd78c2476e16ce9e25b6c024e774ac702"} Nov 25 14:15:03 crc kubenswrapper[4702]: I1125 14:15:03.007943 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401335-2rddm" Nov 25 14:15:03 crc kubenswrapper[4702]: I1125 14:15:03.096659 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npsvm\" (UniqueName: \"kubernetes.io/projected/109f5269-83c5-4ceb-97ab-1eb61a31c4eb-kube-api-access-npsvm\") pod \"109f5269-83c5-4ceb-97ab-1eb61a31c4eb\" (UID: \"109f5269-83c5-4ceb-97ab-1eb61a31c4eb\") " Nov 25 14:15:03 crc kubenswrapper[4702]: I1125 14:15:03.096699 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/109f5269-83c5-4ceb-97ab-1eb61a31c4eb-secret-volume\") pod \"109f5269-83c5-4ceb-97ab-1eb61a31c4eb\" (UID: \"109f5269-83c5-4ceb-97ab-1eb61a31c4eb\") " Nov 25 14:15:03 crc kubenswrapper[4702]: I1125 14:15:03.096729 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/109f5269-83c5-4ceb-97ab-1eb61a31c4eb-config-volume\") pod \"109f5269-83c5-4ceb-97ab-1eb61a31c4eb\" (UID: \"109f5269-83c5-4ceb-97ab-1eb61a31c4eb\") " Nov 25 14:15:03 crc kubenswrapper[4702]: I1125 14:15:03.097929 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/109f5269-83c5-4ceb-97ab-1eb61a31c4eb-config-volume" (OuterVolumeSpecName: "config-volume") pod "109f5269-83c5-4ceb-97ab-1eb61a31c4eb" (UID: "109f5269-83c5-4ceb-97ab-1eb61a31c4eb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 14:15:03 crc kubenswrapper[4702]: I1125 14:15:03.107357 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/109f5269-83c5-4ceb-97ab-1eb61a31c4eb-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "109f5269-83c5-4ceb-97ab-1eb61a31c4eb" (UID: "109f5269-83c5-4ceb-97ab-1eb61a31c4eb"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:15:03 crc kubenswrapper[4702]: I1125 14:15:03.108446 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/109f5269-83c5-4ceb-97ab-1eb61a31c4eb-kube-api-access-npsvm" (OuterVolumeSpecName: "kube-api-access-npsvm") pod "109f5269-83c5-4ceb-97ab-1eb61a31c4eb" (UID: "109f5269-83c5-4ceb-97ab-1eb61a31c4eb"). InnerVolumeSpecName "kube-api-access-npsvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:15:03 crc kubenswrapper[4702]: I1125 14:15:03.199060 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npsvm\" (UniqueName: \"kubernetes.io/projected/109f5269-83c5-4ceb-97ab-1eb61a31c4eb-kube-api-access-npsvm\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:03 crc kubenswrapper[4702]: I1125 14:15:03.199119 4702 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/109f5269-83c5-4ceb-97ab-1eb61a31c4eb-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:03 crc kubenswrapper[4702]: I1125 14:15:03.199135 4702 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/109f5269-83c5-4ceb-97ab-1eb61a31c4eb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:03 crc kubenswrapper[4702]: I1125 14:15:03.755689 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401335-2rddm" event={"ID":"109f5269-83c5-4ceb-97ab-1eb61a31c4eb","Type":"ContainerDied","Data":"00d99b35a275aea79e4f9ceed4c70f9cd78c2476e16ce9e25b6c024e774ac702"} Nov 25 14:15:03 crc kubenswrapper[4702]: I1125 14:15:03.756106 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00d99b35a275aea79e4f9ceed4c70f9cd78c2476e16ce9e25b6c024e774ac702" Nov 25 14:15:03 crc kubenswrapper[4702]: I1125 14:15:03.755780 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401335-2rddm" Nov 25 14:15:10 crc kubenswrapper[4702]: I1125 14:15:10.801557 4702 scope.go:117] "RemoveContainer" containerID="32273717263d7fcc99e8193a439dd976da66869c55bb4633d9314cacb6ab92a1" Nov 25 14:15:10 crc kubenswrapper[4702]: E1125 14:15:10.802572 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qqr76_openshift-machine-config-operator(3c491818-b0bb-4d82-a031-96e2dfac8c27)\"" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" Nov 25 14:15:23 crc kubenswrapper[4702]: I1125 14:15:23.801925 4702 scope.go:117] "RemoveContainer" containerID="32273717263d7fcc99e8193a439dd976da66869c55bb4633d9314cacb6ab92a1" Nov 25 14:15:23 crc kubenswrapper[4702]: E1125 14:15:23.803415 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qqr76_openshift-machine-config-operator(3c491818-b0bb-4d82-a031-96e2dfac8c27)\"" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" Nov 25 14:15:32 crc kubenswrapper[4702]: I1125 14:15:32.386929 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 14:15:32 crc kubenswrapper[4702]: I1125 14:15:32.387839 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="2b5b9839-7217-4d14-a6e3-c8e62579f40e" containerName="glance-log" containerID="cri-o://c91838e0db075b4ce25a4f519b9aa59e579ebf0f2e0c6fd0aadaf23dae87e9c6" gracePeriod=30 Nov 25 14:15:32 crc kubenswrapper[4702]: I1125 14:15:32.387887 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="2b5b9839-7217-4d14-a6e3-c8e62579f40e" containerName="glance-api" containerID="cri-o://3e26d313616b3a70fe756a3d41215ec16acd28c473c3365fe61988d4e73729a1" gracePeriod=30 Nov 25 14:15:32 crc kubenswrapper[4702]: I1125 14:15:32.387970 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="2b5b9839-7217-4d14-a6e3-c8e62579f40e" containerName="glance-httpd" containerID="cri-o://178cd9604dee844af6359851c1ff829c6dc3b24c42e2b65fd64e216a696ca23d" gracePeriod=30 Nov 25 14:15:32 crc kubenswrapper[4702]: I1125 14:15:32.997105 4702 generic.go:334] "Generic (PLEG): container finished" podID="2b5b9839-7217-4d14-a6e3-c8e62579f40e" containerID="3e26d313616b3a70fe756a3d41215ec16acd28c473c3365fe61988d4e73729a1" exitCode=0 Nov 25 14:15:32 crc kubenswrapper[4702]: I1125 14:15:32.997442 4702 generic.go:334] "Generic (PLEG): container finished" podID="2b5b9839-7217-4d14-a6e3-c8e62579f40e" containerID="178cd9604dee844af6359851c1ff829c6dc3b24c42e2b65fd64e216a696ca23d" exitCode=0 Nov 25 14:15:32 crc kubenswrapper[4702]: I1125 14:15:32.997148 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"2b5b9839-7217-4d14-a6e3-c8e62579f40e","Type":"ContainerDied","Data":"3e26d313616b3a70fe756a3d41215ec16acd28c473c3365fe61988d4e73729a1"} Nov 25 14:15:32 crc kubenswrapper[4702]: I1125 14:15:32.997485 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"2b5b9839-7217-4d14-a6e3-c8e62579f40e","Type":"ContainerDied","Data":"178cd9604dee844af6359851c1ff829c6dc3b24c42e2b65fd64e216a696ca23d"} Nov 25 14:15:32 crc kubenswrapper[4702]: I1125 14:15:32.997500 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"2b5b9839-7217-4d14-a6e3-c8e62579f40e","Type":"ContainerDied","Data":"c91838e0db075b4ce25a4f519b9aa59e579ebf0f2e0c6fd0aadaf23dae87e9c6"} Nov 25 14:15:32 crc kubenswrapper[4702]: I1125 14:15:32.997456 4702 generic.go:334] "Generic (PLEG): container finished" podID="2b5b9839-7217-4d14-a6e3-c8e62579f40e" containerID="c91838e0db075b4ce25a4f519b9aa59e579ebf0f2e0c6fd0aadaf23dae87e9c6" exitCode=143 Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.382740 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.504631 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lfn2\" (UniqueName: \"kubernetes.io/projected/2b5b9839-7217-4d14-a6e3-c8e62579f40e-kube-api-access-4lfn2\") pod \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.504810 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-lib-modules\") pod \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.504908 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-run\") pod \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.504907 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "2b5b9839-7217-4d14-a6e3-c8e62579f40e" (UID: "2b5b9839-7217-4d14-a6e3-c8e62579f40e"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.504979 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-run" (OuterVolumeSpecName: "run") pod "2b5b9839-7217-4d14-a6e3-c8e62579f40e" (UID: "2b5b9839-7217-4d14-a6e3-c8e62579f40e"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.505026 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-sys\") pod \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.505047 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-sys" (OuterVolumeSpecName: "sys") pod "2b5b9839-7217-4d14-a6e3-c8e62579f40e" (UID: "2b5b9839-7217-4d14-a6e3-c8e62579f40e"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.505056 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-etc-iscsi\") pod \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.505086 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b5b9839-7217-4d14-a6e3-c8e62579f40e-config-data\") pod \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.505111 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.505138 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "2b5b9839-7217-4d14-a6e3-c8e62579f40e" (UID: "2b5b9839-7217-4d14-a6e3-c8e62579f40e"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.505146 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-dev\") pod \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.505166 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-dev" (OuterVolumeSpecName: "dev") pod "2b5b9839-7217-4d14-a6e3-c8e62579f40e" (UID: "2b5b9839-7217-4d14-a6e3-c8e62579f40e"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.505178 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b5b9839-7217-4d14-a6e3-c8e62579f40e-scripts\") pod \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.505518 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-var-locks-brick\") pod \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.505552 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b5b9839-7217-4d14-a6e3-c8e62579f40e-logs\") pod \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.505580 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b5b9839-7217-4d14-a6e3-c8e62579f40e-httpd-run\") pod \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.505609 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.505629 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-etc-nvme\") pod \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\" (UID: \"2b5b9839-7217-4d14-a6e3-c8e62579f40e\") " Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.506114 4702 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.506133 4702 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-run\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.506256 4702 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-sys\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.506271 4702 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.506282 4702 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-dev\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.506197 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b5b9839-7217-4d14-a6e3-c8e62579f40e-logs" (OuterVolumeSpecName: "logs") pod "2b5b9839-7217-4d14-a6e3-c8e62579f40e" (UID: "2b5b9839-7217-4d14-a6e3-c8e62579f40e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.506299 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "2b5b9839-7217-4d14-a6e3-c8e62579f40e" (UID: "2b5b9839-7217-4d14-a6e3-c8e62579f40e"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.506222 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "2b5b9839-7217-4d14-a6e3-c8e62579f40e" (UID: "2b5b9839-7217-4d14-a6e3-c8e62579f40e"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.506543 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b5b9839-7217-4d14-a6e3-c8e62579f40e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2b5b9839-7217-4d14-a6e3-c8e62579f40e" (UID: "2b5b9839-7217-4d14-a6e3-c8e62579f40e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.511877 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "2b5b9839-7217-4d14-a6e3-c8e62579f40e" (UID: "2b5b9839-7217-4d14-a6e3-c8e62579f40e"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.512217 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b5b9839-7217-4d14-a6e3-c8e62579f40e-scripts" (OuterVolumeSpecName: "scripts") pod "2b5b9839-7217-4d14-a6e3-c8e62579f40e" (UID: "2b5b9839-7217-4d14-a6e3-c8e62579f40e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.514310 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance-cache") pod "2b5b9839-7217-4d14-a6e3-c8e62579f40e" (UID: "2b5b9839-7217-4d14-a6e3-c8e62579f40e"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.514427 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b5b9839-7217-4d14-a6e3-c8e62579f40e-kube-api-access-4lfn2" (OuterVolumeSpecName: "kube-api-access-4lfn2") pod "2b5b9839-7217-4d14-a6e3-c8e62579f40e" (UID: "2b5b9839-7217-4d14-a6e3-c8e62579f40e"). InnerVolumeSpecName "kube-api-access-4lfn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.569539 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b5b9839-7217-4d14-a6e3-c8e62579f40e-config-data" (OuterVolumeSpecName: "config-data") pod "2b5b9839-7217-4d14-a6e3-c8e62579f40e" (UID: "2b5b9839-7217-4d14-a6e3-c8e62579f40e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.607949 4702 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b5b9839-7217-4d14-a6e3-c8e62579f40e-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.608012 4702 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.608023 4702 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b5b9839-7217-4d14-a6e3-c8e62579f40e-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.608033 4702 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.608044 4702 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b5b9839-7217-4d14-a6e3-c8e62579f40e-logs\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.608052 4702 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b5b9839-7217-4d14-a6e3-c8e62579f40e-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.608066 4702 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.608075 4702 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2b5b9839-7217-4d14-a6e3-c8e62579f40e-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.608092 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lfn2\" (UniqueName: \"kubernetes.io/projected/2b5b9839-7217-4d14-a6e3-c8e62579f40e-kube-api-access-4lfn2\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.620894 4702 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.621103 4702 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.710416 4702 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.710469 4702 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.791319 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-zmh5f"] Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.797726 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-zmh5f"] Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.809226 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf33d671-2bc3-46aa-95e9-eceb37bfe864" path="/var/lib/kubelet/pods/cf33d671-2bc3-46aa-95e9-eceb37bfe864/volumes" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.845395 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance9a19-account-delete-xggfm"] Nov 25 14:15:33 crc kubenswrapper[4702]: E1125 14:15:33.845752 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="109f5269-83c5-4ceb-97ab-1eb61a31c4eb" containerName="collect-profiles" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.845768 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="109f5269-83c5-4ceb-97ab-1eb61a31c4eb" containerName="collect-profiles" Nov 25 14:15:33 crc kubenswrapper[4702]: E1125 14:15:33.845784 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b5b9839-7217-4d14-a6e3-c8e62579f40e" containerName="glance-log" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.845790 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b5b9839-7217-4d14-a6e3-c8e62579f40e" containerName="glance-log" Nov 25 14:15:33 crc kubenswrapper[4702]: E1125 14:15:33.845805 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b5b9839-7217-4d14-a6e3-c8e62579f40e" containerName="glance-httpd" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.845812 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b5b9839-7217-4d14-a6e3-c8e62579f40e" containerName="glance-httpd" Nov 25 14:15:33 crc kubenswrapper[4702]: E1125 14:15:33.845824 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b5b9839-7217-4d14-a6e3-c8e62579f40e" containerName="glance-api" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.845829 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b5b9839-7217-4d14-a6e3-c8e62579f40e" containerName="glance-api" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.845976 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b5b9839-7217-4d14-a6e3-c8e62579f40e" containerName="glance-log" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.845990 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b5b9839-7217-4d14-a6e3-c8e62579f40e" containerName="glance-httpd" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.845999 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b5b9839-7217-4d14-a6e3-c8e62579f40e" containerName="glance-api" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.846009 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="109f5269-83c5-4ceb-97ab-1eb61a31c4eb" containerName="collect-profiles" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.846536 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance9a19-account-delete-xggfm" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.858672 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance9a19-account-delete-xggfm"] Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.913413 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4-operator-scripts\") pod \"glance9a19-account-delete-xggfm\" (UID: \"e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4\") " pod="glance-kuttl-tests/glance9a19-account-delete-xggfm" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.913804 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsdj9\" (UniqueName: \"kubernetes.io/projected/e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4-kube-api-access-hsdj9\") pod \"glance9a19-account-delete-xggfm\" (UID: \"e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4\") " pod="glance-kuttl-tests/glance9a19-account-delete-xggfm" Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.931993 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.932366 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="50e365be-ffbe-4bf9-8eec-19cca5ca20cb" containerName="glance-log" containerID="cri-o://cd9ebd18fca903bcc5c3910acd17c494abdac265821f0092b48049e2d699a783" gracePeriod=30 Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.932466 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="50e365be-ffbe-4bf9-8eec-19cca5ca20cb" containerName="glance-api" containerID="cri-o://5df1dfc91583f02a0504817678e09b0d330f585b91b0df332270cca5ebc47573" gracePeriod=30 Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.932633 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="50e365be-ffbe-4bf9-8eec-19cca5ca20cb" containerName="glance-httpd" containerID="cri-o://f6440decba611fb6665099ea729c5d8b60b42b55644f2be04030b74123ceead2" gracePeriod=30 Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.992488 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.993064 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="2ba715ae-64e0-4810-b14d-f0c73b02f65d" containerName="glance-log" containerID="cri-o://094e6d8e70f0a27cf92b867a94892cc0e5547cae1c180d5a2e21ad71123c29ec" gracePeriod=30 Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.993309 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="2ba715ae-64e0-4810-b14d-f0c73b02f65d" containerName="glance-api" containerID="cri-o://4cf74f8aeeaa938a660d0ca5bfdef3ebfaebf66bb0ef971a12b7ef0b5dda5e11" gracePeriod=30 Nov 25 14:15:33 crc kubenswrapper[4702]: I1125 14:15:33.993442 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="2ba715ae-64e0-4810-b14d-f0c73b02f65d" containerName="glance-httpd" containerID="cri-o://b07b7cb2e145043a8a4f4070f63ce4ef540385d4bdc101570ce60ffaf6ac41b1" gracePeriod=30 Nov 25 14:15:34 crc kubenswrapper[4702]: I1125 14:15:34.015342 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4-operator-scripts\") pod \"glance9a19-account-delete-xggfm\" (UID: \"e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4\") " pod="glance-kuttl-tests/glance9a19-account-delete-xggfm" Nov 25 14:15:34 crc kubenswrapper[4702]: I1125 14:15:34.016125 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsdj9\" (UniqueName: \"kubernetes.io/projected/e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4-kube-api-access-hsdj9\") pod \"glance9a19-account-delete-xggfm\" (UID: \"e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4\") " pod="glance-kuttl-tests/glance9a19-account-delete-xggfm" Nov 25 14:15:34 crc kubenswrapper[4702]: I1125 14:15:34.016415 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"2b5b9839-7217-4d14-a6e3-c8e62579f40e","Type":"ContainerDied","Data":"bae79b726b36329b7fbf6dae1150e950506789af04b0c406f4327ea2b7bf7732"} Nov 25 14:15:34 crc kubenswrapper[4702]: I1125 14:15:34.016188 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4-operator-scripts\") pod \"glance9a19-account-delete-xggfm\" (UID: \"e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4\") " pod="glance-kuttl-tests/glance9a19-account-delete-xggfm" Nov 25 14:15:34 crc kubenswrapper[4702]: I1125 14:15:34.016503 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 25 14:15:34 crc kubenswrapper[4702]: I1125 14:15:34.016566 4702 scope.go:117] "RemoveContainer" containerID="3e26d313616b3a70fe756a3d41215ec16acd28c473c3365fe61988d4e73729a1" Nov 25 14:15:34 crc kubenswrapper[4702]: I1125 14:15:34.042520 4702 scope.go:117] "RemoveContainer" containerID="178cd9604dee844af6359851c1ff829c6dc3b24c42e2b65fd64e216a696ca23d" Nov 25 14:15:34 crc kubenswrapper[4702]: I1125 14:15:34.050558 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsdj9\" (UniqueName: \"kubernetes.io/projected/e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4-kube-api-access-hsdj9\") pod \"glance9a19-account-delete-xggfm\" (UID: \"e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4\") " pod="glance-kuttl-tests/glance9a19-account-delete-xggfm" Nov 25 14:15:34 crc kubenswrapper[4702]: I1125 14:15:34.053359 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 14:15:34 crc kubenswrapper[4702]: I1125 14:15:34.065076 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 25 14:15:34 crc kubenswrapper[4702]: I1125 14:15:34.135520 4702 scope.go:117] "RemoveContainer" containerID="c91838e0db075b4ce25a4f519b9aa59e579ebf0f2e0c6fd0aadaf23dae87e9c6" Nov 25 14:15:34 crc kubenswrapper[4702]: I1125 14:15:34.200864 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance9a19-account-delete-xggfm" Nov 25 14:15:34 crc kubenswrapper[4702]: I1125 14:15:34.652820 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance9a19-account-delete-xggfm"] Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.008269 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.027548 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.028317 4702 generic.go:334] "Generic (PLEG): container finished" podID="2ba715ae-64e0-4810-b14d-f0c73b02f65d" containerID="4cf74f8aeeaa938a660d0ca5bfdef3ebfaebf66bb0ef971a12b7ef0b5dda5e11" exitCode=0 Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.028345 4702 generic.go:334] "Generic (PLEG): container finished" podID="2ba715ae-64e0-4810-b14d-f0c73b02f65d" containerID="b07b7cb2e145043a8a4f4070f63ce4ef540385d4bdc101570ce60ffaf6ac41b1" exitCode=0 Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.028355 4702 generic.go:334] "Generic (PLEG): container finished" podID="2ba715ae-64e0-4810-b14d-f0c73b02f65d" containerID="094e6d8e70f0a27cf92b867a94892cc0e5547cae1c180d5a2e21ad71123c29ec" exitCode=143 Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.028351 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"2ba715ae-64e0-4810-b14d-f0c73b02f65d","Type":"ContainerDied","Data":"4cf74f8aeeaa938a660d0ca5bfdef3ebfaebf66bb0ef971a12b7ef0b5dda5e11"} Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.028390 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"2ba715ae-64e0-4810-b14d-f0c73b02f65d","Type":"ContainerDied","Data":"b07b7cb2e145043a8a4f4070f63ce4ef540385d4bdc101570ce60ffaf6ac41b1"} Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.028404 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"2ba715ae-64e0-4810-b14d-f0c73b02f65d","Type":"ContainerDied","Data":"094e6d8e70f0a27cf92b867a94892cc0e5547cae1c180d5a2e21ad71123c29ec"} Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.028413 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"2ba715ae-64e0-4810-b14d-f0c73b02f65d","Type":"ContainerDied","Data":"4c1fd19dc3bd9fc157489944a459533d90c6ee14297b20c83e64ee525e00ebde"} Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.028415 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.028440 4702 scope.go:117] "RemoveContainer" containerID="4cf74f8aeeaa938a660d0ca5bfdef3ebfaebf66bb0ef971a12b7ef0b5dda5e11" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.030796 4702 generic.go:334] "Generic (PLEG): container finished" podID="50e365be-ffbe-4bf9-8eec-19cca5ca20cb" containerID="5df1dfc91583f02a0504817678e09b0d330f585b91b0df332270cca5ebc47573" exitCode=0 Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.030816 4702 generic.go:334] "Generic (PLEG): container finished" podID="50e365be-ffbe-4bf9-8eec-19cca5ca20cb" containerID="f6440decba611fb6665099ea729c5d8b60b42b55644f2be04030b74123ceead2" exitCode=0 Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.030824 4702 generic.go:334] "Generic (PLEG): container finished" podID="50e365be-ffbe-4bf9-8eec-19cca5ca20cb" containerID="cd9ebd18fca903bcc5c3910acd17c494abdac265821f0092b48049e2d699a783" exitCode=143 Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.030860 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"50e365be-ffbe-4bf9-8eec-19cca5ca20cb","Type":"ContainerDied","Data":"5df1dfc91583f02a0504817678e09b0d330f585b91b0df332270cca5ebc47573"} Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.030878 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"50e365be-ffbe-4bf9-8eec-19cca5ca20cb","Type":"ContainerDied","Data":"f6440decba611fb6665099ea729c5d8b60b42b55644f2be04030b74123ceead2"} Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.030889 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"50e365be-ffbe-4bf9-8eec-19cca5ca20cb","Type":"ContainerDied","Data":"cd9ebd18fca903bcc5c3910acd17c494abdac265821f0092b48049e2d699a783"} Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.030898 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"50e365be-ffbe-4bf9-8eec-19cca5ca20cb","Type":"ContainerDied","Data":"fac158d959802d93193f8df09824e8bb237100e63101313bd6087057449c101b"} Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.030952 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.032534 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance9a19-account-delete-xggfm" event={"ID":"e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4","Type":"ContainerStarted","Data":"0536a3dac6b639535a8e04cbcbdb471a210f9dd77cf0b1d4391a71227e1d1e17"} Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.032562 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance9a19-account-delete-xggfm" event={"ID":"e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4","Type":"ContainerStarted","Data":"a8a04f149a6f55bf9228ad076b5fd5dcf76e0299f393d6b3fba6ee6f38524ca7"} Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.048663 4702 scope.go:117] "RemoveContainer" containerID="b07b7cb2e145043a8a4f4070f63ce4ef540385d4bdc101570ce60ffaf6ac41b1" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.075051 4702 scope.go:117] "RemoveContainer" containerID="094e6d8e70f0a27cf92b867a94892cc0e5547cae1c180d5a2e21ad71123c29ec" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.097021 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance9a19-account-delete-xggfm" podStartSLOduration=2.096985592 podStartE2EDuration="2.096985592s" podCreationTimestamp="2025-11-25 14:15:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 14:15:35.093551376 +0000 UTC m=+1744.256201826" watchObservedRunningTime="2025-11-25 14:15:35.096985592 +0000 UTC m=+1744.259636042" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.098018 4702 scope.go:117] "RemoveContainer" containerID="4cf74f8aeeaa938a660d0ca5bfdef3ebfaebf66bb0ef971a12b7ef0b5dda5e11" Nov 25 14:15:35 crc kubenswrapper[4702]: E1125 14:15:35.099037 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cf74f8aeeaa938a660d0ca5bfdef3ebfaebf66bb0ef971a12b7ef0b5dda5e11\": container with ID starting with 4cf74f8aeeaa938a660d0ca5bfdef3ebfaebf66bb0ef971a12b7ef0b5dda5e11 not found: ID does not exist" containerID="4cf74f8aeeaa938a660d0ca5bfdef3ebfaebf66bb0ef971a12b7ef0b5dda5e11" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.099076 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cf74f8aeeaa938a660d0ca5bfdef3ebfaebf66bb0ef971a12b7ef0b5dda5e11"} err="failed to get container status \"4cf74f8aeeaa938a660d0ca5bfdef3ebfaebf66bb0ef971a12b7ef0b5dda5e11\": rpc error: code = NotFound desc = could not find container \"4cf74f8aeeaa938a660d0ca5bfdef3ebfaebf66bb0ef971a12b7ef0b5dda5e11\": container with ID starting with 4cf74f8aeeaa938a660d0ca5bfdef3ebfaebf66bb0ef971a12b7ef0b5dda5e11 not found: ID does not exist" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.099104 4702 scope.go:117] "RemoveContainer" containerID="b07b7cb2e145043a8a4f4070f63ce4ef540385d4bdc101570ce60ffaf6ac41b1" Nov 25 14:15:35 crc kubenswrapper[4702]: E1125 14:15:35.099596 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b07b7cb2e145043a8a4f4070f63ce4ef540385d4bdc101570ce60ffaf6ac41b1\": container with ID starting with b07b7cb2e145043a8a4f4070f63ce4ef540385d4bdc101570ce60ffaf6ac41b1 not found: ID does not exist" containerID="b07b7cb2e145043a8a4f4070f63ce4ef540385d4bdc101570ce60ffaf6ac41b1" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.099644 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b07b7cb2e145043a8a4f4070f63ce4ef540385d4bdc101570ce60ffaf6ac41b1"} err="failed to get container status \"b07b7cb2e145043a8a4f4070f63ce4ef540385d4bdc101570ce60ffaf6ac41b1\": rpc error: code = NotFound desc = could not find container \"b07b7cb2e145043a8a4f4070f63ce4ef540385d4bdc101570ce60ffaf6ac41b1\": container with ID starting with b07b7cb2e145043a8a4f4070f63ce4ef540385d4bdc101570ce60ffaf6ac41b1 not found: ID does not exist" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.099676 4702 scope.go:117] "RemoveContainer" containerID="094e6d8e70f0a27cf92b867a94892cc0e5547cae1c180d5a2e21ad71123c29ec" Nov 25 14:15:35 crc kubenswrapper[4702]: E1125 14:15:35.100067 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"094e6d8e70f0a27cf92b867a94892cc0e5547cae1c180d5a2e21ad71123c29ec\": container with ID starting with 094e6d8e70f0a27cf92b867a94892cc0e5547cae1c180d5a2e21ad71123c29ec not found: ID does not exist" containerID="094e6d8e70f0a27cf92b867a94892cc0e5547cae1c180d5a2e21ad71123c29ec" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.100110 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"094e6d8e70f0a27cf92b867a94892cc0e5547cae1c180d5a2e21ad71123c29ec"} err="failed to get container status \"094e6d8e70f0a27cf92b867a94892cc0e5547cae1c180d5a2e21ad71123c29ec\": rpc error: code = NotFound desc = could not find container \"094e6d8e70f0a27cf92b867a94892cc0e5547cae1c180d5a2e21ad71123c29ec\": container with ID starting with 094e6d8e70f0a27cf92b867a94892cc0e5547cae1c180d5a2e21ad71123c29ec not found: ID does not exist" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.100131 4702 scope.go:117] "RemoveContainer" containerID="4cf74f8aeeaa938a660d0ca5bfdef3ebfaebf66bb0ef971a12b7ef0b5dda5e11" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.100477 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cf74f8aeeaa938a660d0ca5bfdef3ebfaebf66bb0ef971a12b7ef0b5dda5e11"} err="failed to get container status \"4cf74f8aeeaa938a660d0ca5bfdef3ebfaebf66bb0ef971a12b7ef0b5dda5e11\": rpc error: code = NotFound desc = could not find container \"4cf74f8aeeaa938a660d0ca5bfdef3ebfaebf66bb0ef971a12b7ef0b5dda5e11\": container with ID starting with 4cf74f8aeeaa938a660d0ca5bfdef3ebfaebf66bb0ef971a12b7ef0b5dda5e11 not found: ID does not exist" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.100551 4702 scope.go:117] "RemoveContainer" containerID="b07b7cb2e145043a8a4f4070f63ce4ef540385d4bdc101570ce60ffaf6ac41b1" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.100878 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b07b7cb2e145043a8a4f4070f63ce4ef540385d4bdc101570ce60ffaf6ac41b1"} err="failed to get container status \"b07b7cb2e145043a8a4f4070f63ce4ef540385d4bdc101570ce60ffaf6ac41b1\": rpc error: code = NotFound desc = could not find container \"b07b7cb2e145043a8a4f4070f63ce4ef540385d4bdc101570ce60ffaf6ac41b1\": container with ID starting with b07b7cb2e145043a8a4f4070f63ce4ef540385d4bdc101570ce60ffaf6ac41b1 not found: ID does not exist" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.100897 4702 scope.go:117] "RemoveContainer" containerID="094e6d8e70f0a27cf92b867a94892cc0e5547cae1c180d5a2e21ad71123c29ec" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.101168 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"094e6d8e70f0a27cf92b867a94892cc0e5547cae1c180d5a2e21ad71123c29ec"} err="failed to get container status \"094e6d8e70f0a27cf92b867a94892cc0e5547cae1c180d5a2e21ad71123c29ec\": rpc error: code = NotFound desc = could not find container \"094e6d8e70f0a27cf92b867a94892cc0e5547cae1c180d5a2e21ad71123c29ec\": container with ID starting with 094e6d8e70f0a27cf92b867a94892cc0e5547cae1c180d5a2e21ad71123c29ec not found: ID does not exist" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.101192 4702 scope.go:117] "RemoveContainer" containerID="4cf74f8aeeaa938a660d0ca5bfdef3ebfaebf66bb0ef971a12b7ef0b5dda5e11" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.101736 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cf74f8aeeaa938a660d0ca5bfdef3ebfaebf66bb0ef971a12b7ef0b5dda5e11"} err="failed to get container status \"4cf74f8aeeaa938a660d0ca5bfdef3ebfaebf66bb0ef971a12b7ef0b5dda5e11\": rpc error: code = NotFound desc = could not find container \"4cf74f8aeeaa938a660d0ca5bfdef3ebfaebf66bb0ef971a12b7ef0b5dda5e11\": container with ID starting with 4cf74f8aeeaa938a660d0ca5bfdef3ebfaebf66bb0ef971a12b7ef0b5dda5e11 not found: ID does not exist" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.101758 4702 scope.go:117] "RemoveContainer" containerID="b07b7cb2e145043a8a4f4070f63ce4ef540385d4bdc101570ce60ffaf6ac41b1" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.101993 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b07b7cb2e145043a8a4f4070f63ce4ef540385d4bdc101570ce60ffaf6ac41b1"} err="failed to get container status \"b07b7cb2e145043a8a4f4070f63ce4ef540385d4bdc101570ce60ffaf6ac41b1\": rpc error: code = NotFound desc = could not find container \"b07b7cb2e145043a8a4f4070f63ce4ef540385d4bdc101570ce60ffaf6ac41b1\": container with ID starting with b07b7cb2e145043a8a4f4070f63ce4ef540385d4bdc101570ce60ffaf6ac41b1 not found: ID does not exist" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.102018 4702 scope.go:117] "RemoveContainer" containerID="094e6d8e70f0a27cf92b867a94892cc0e5547cae1c180d5a2e21ad71123c29ec" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.102403 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"094e6d8e70f0a27cf92b867a94892cc0e5547cae1c180d5a2e21ad71123c29ec"} err="failed to get container status \"094e6d8e70f0a27cf92b867a94892cc0e5547cae1c180d5a2e21ad71123c29ec\": rpc error: code = NotFound desc = could not find container \"094e6d8e70f0a27cf92b867a94892cc0e5547cae1c180d5a2e21ad71123c29ec\": container with ID starting with 094e6d8e70f0a27cf92b867a94892cc0e5547cae1c180d5a2e21ad71123c29ec not found: ID does not exist" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.102426 4702 scope.go:117] "RemoveContainer" containerID="5df1dfc91583f02a0504817678e09b0d330f585b91b0df332270cca5ebc47573" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.120622 4702 scope.go:117] "RemoveContainer" containerID="f6440decba611fb6665099ea729c5d8b60b42b55644f2be04030b74123ceead2" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.136687 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-var-locks-brick\") pod \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.136733 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-httpd-run\") pod \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.136759 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-etc-nvme\") pod \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.136789 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ba715ae-64e0-4810-b14d-f0c73b02f65d-config-data\") pod \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.136802 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-etc-iscsi\") pod \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.136822 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-var-locks-brick\") pod \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.136870 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-etc-nvme\") pod \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.136894 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vcd6s\" (UniqueName: \"kubernetes.io/projected/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-kube-api-access-vcd6s\") pod \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.136914 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-run\") pod \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.136934 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-lib-modules\") pod \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.136904 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "50e365be-ffbe-4bf9-8eec-19cca5ca20cb" (UID: "50e365be-ffbe-4bf9-8eec-19cca5ca20cb"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.136943 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "2ba715ae-64e0-4810-b14d-f0c73b02f65d" (UID: "2ba715ae-64e0-4810-b14d-f0c73b02f65d"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.136954 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbxxm\" (UniqueName: \"kubernetes.io/projected/2ba715ae-64e0-4810-b14d-f0c73b02f65d-kube-api-access-rbxxm\") pod \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.137056 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.137138 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.137172 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-logs\") pod \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.137221 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ba715ae-64e0-4810-b14d-f0c73b02f65d-scripts\") pod \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.137334 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ba715ae-64e0-4810-b14d-f0c73b02f65d-logs\") pod \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.137367 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-sys\") pod \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.137427 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-lib-modules\") pod \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.137427 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "50e365be-ffbe-4bf9-8eec-19cca5ca20cb" (UID: "50e365be-ffbe-4bf9-8eec-19cca5ca20cb"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.137460 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-dev\") pod \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.137491 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-dev" (OuterVolumeSpecName: "dev") pod "2ba715ae-64e0-4810-b14d-f0c73b02f65d" (UID: "2ba715ae-64e0-4810-b14d-f0c73b02f65d"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.137555 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-sys" (OuterVolumeSpecName: "sys") pod "2ba715ae-64e0-4810-b14d-f0c73b02f65d" (UID: "2ba715ae-64e0-4810-b14d-f0c73b02f65d"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.137588 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "2ba715ae-64e0-4810-b14d-f0c73b02f65d" (UID: "2ba715ae-64e0-4810-b14d-f0c73b02f65d"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.137616 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-run" (OuterVolumeSpecName: "run") pod "2ba715ae-64e0-4810-b14d-f0c73b02f65d" (UID: "2ba715ae-64e0-4810-b14d-f0c73b02f65d"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.137704 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "50e365be-ffbe-4bf9-8eec-19cca5ca20cb" (UID: "50e365be-ffbe-4bf9-8eec-19cca5ca20cb"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.137728 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "2ba715ae-64e0-4810-b14d-f0c73b02f65d" (UID: "2ba715ae-64e0-4810-b14d-f0c73b02f65d"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.137757 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "2ba715ae-64e0-4810-b14d-f0c73b02f65d" (UID: "2ba715ae-64e0-4810-b14d-f0c73b02f65d"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.138026 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.138062 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-run\") pod \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.138078 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-dev\") pod \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.138110 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "50e365be-ffbe-4bf9-8eec-19cca5ca20cb" (UID: "50e365be-ffbe-4bf9-8eec-19cca5ca20cb"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.138141 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2ba715ae-64e0-4810-b14d-f0c73b02f65d-httpd-run\") pod \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\" (UID: \"2ba715ae-64e0-4810-b14d-f0c73b02f65d\") " Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.138153 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-run" (OuterVolumeSpecName: "run") pod "50e365be-ffbe-4bf9-8eec-19cca5ca20cb" (UID: "50e365be-ffbe-4bf9-8eec-19cca5ca20cb"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.138161 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-dev" (OuterVolumeSpecName: "dev") pod "50e365be-ffbe-4bf9-8eec-19cca5ca20cb" (UID: "50e365be-ffbe-4bf9-8eec-19cca5ca20cb"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.138176 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-config-data\") pod \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.138217 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-scripts\") pod \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.138284 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-etc-iscsi\") pod \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.138346 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.138363 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-sys\") pod \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\" (UID: \"50e365be-ffbe-4bf9-8eec-19cca5ca20cb\") " Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.138426 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ba715ae-64e0-4810-b14d-f0c73b02f65d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2ba715ae-64e0-4810-b14d-f0c73b02f65d" (UID: "2ba715ae-64e0-4810-b14d-f0c73b02f65d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.138454 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ba715ae-64e0-4810-b14d-f0c73b02f65d-logs" (OuterVolumeSpecName: "logs") pod "2ba715ae-64e0-4810-b14d-f0c73b02f65d" (UID: "2ba715ae-64e0-4810-b14d-f0c73b02f65d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.139033 4702 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.139046 4702 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.139057 4702 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.139067 4702 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.139076 4702 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.139073 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-logs" (OuterVolumeSpecName: "logs") pod "50e365be-ffbe-4bf9-8eec-19cca5ca20cb" (UID: "50e365be-ffbe-4bf9-8eec-19cca5ca20cb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.139086 4702 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.139096 4702 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-run\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.139105 4702 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.139114 4702 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ba715ae-64e0-4810-b14d-f0c73b02f65d-logs\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.139113 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "50e365be-ffbe-4bf9-8eec-19cca5ca20cb" (UID: "50e365be-ffbe-4bf9-8eec-19cca5ca20cb"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.139122 4702 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-sys\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.139133 4702 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.139142 4702 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2ba715ae-64e0-4810-b14d-f0c73b02f65d-dev\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.139153 4702 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-run\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.139161 4702 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-dev\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.139169 4702 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2ba715ae-64e0-4810-b14d-f0c73b02f65d-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.139142 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-sys" (OuterVolumeSpecName: "sys") pod "50e365be-ffbe-4bf9-8eec-19cca5ca20cb" (UID: "50e365be-ffbe-4bf9-8eec-19cca5ca20cb"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.142105 4702 scope.go:117] "RemoveContainer" containerID="cd9ebd18fca903bcc5c3910acd17c494abdac265821f0092b48049e2d699a783" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.144201 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance-cache") pod "2ba715ae-64e0-4810-b14d-f0c73b02f65d" (UID: "2ba715ae-64e0-4810-b14d-f0c73b02f65d"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.144307 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "2ba715ae-64e0-4810-b14d-f0c73b02f65d" (UID: "2ba715ae-64e0-4810-b14d-f0c73b02f65d"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.144706 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "50e365be-ffbe-4bf9-8eec-19cca5ca20cb" (UID: "50e365be-ffbe-4bf9-8eec-19cca5ca20cb"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.145097 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ba715ae-64e0-4810-b14d-f0c73b02f65d-scripts" (OuterVolumeSpecName: "scripts") pod "2ba715ae-64e0-4810-b14d-f0c73b02f65d" (UID: "2ba715ae-64e0-4810-b14d-f0c73b02f65d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.145325 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-kube-api-access-vcd6s" (OuterVolumeSpecName: "kube-api-access-vcd6s") pod "50e365be-ffbe-4bf9-8eec-19cca5ca20cb" (UID: "50e365be-ffbe-4bf9-8eec-19cca5ca20cb"). InnerVolumeSpecName "kube-api-access-vcd6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.145627 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ba715ae-64e0-4810-b14d-f0c73b02f65d-kube-api-access-rbxxm" (OuterVolumeSpecName: "kube-api-access-rbxxm") pod "2ba715ae-64e0-4810-b14d-f0c73b02f65d" (UID: "2ba715ae-64e0-4810-b14d-f0c73b02f65d"). InnerVolumeSpecName "kube-api-access-rbxxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.146050 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance-cache") pod "50e365be-ffbe-4bf9-8eec-19cca5ca20cb" (UID: "50e365be-ffbe-4bf9-8eec-19cca5ca20cb"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.146550 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-scripts" (OuterVolumeSpecName: "scripts") pod "50e365be-ffbe-4bf9-8eec-19cca5ca20cb" (UID: "50e365be-ffbe-4bf9-8eec-19cca5ca20cb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.173165 4702 scope.go:117] "RemoveContainer" containerID="5df1dfc91583f02a0504817678e09b0d330f585b91b0df332270cca5ebc47573" Nov 25 14:15:35 crc kubenswrapper[4702]: E1125 14:15:35.174021 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5df1dfc91583f02a0504817678e09b0d330f585b91b0df332270cca5ebc47573\": container with ID starting with 5df1dfc91583f02a0504817678e09b0d330f585b91b0df332270cca5ebc47573 not found: ID does not exist" containerID="5df1dfc91583f02a0504817678e09b0d330f585b91b0df332270cca5ebc47573" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.174078 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5df1dfc91583f02a0504817678e09b0d330f585b91b0df332270cca5ebc47573"} err="failed to get container status \"5df1dfc91583f02a0504817678e09b0d330f585b91b0df332270cca5ebc47573\": rpc error: code = NotFound desc = could not find container \"5df1dfc91583f02a0504817678e09b0d330f585b91b0df332270cca5ebc47573\": container with ID starting with 5df1dfc91583f02a0504817678e09b0d330f585b91b0df332270cca5ebc47573 not found: ID does not exist" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.174141 4702 scope.go:117] "RemoveContainer" containerID="f6440decba611fb6665099ea729c5d8b60b42b55644f2be04030b74123ceead2" Nov 25 14:15:35 crc kubenswrapper[4702]: E1125 14:15:35.174718 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6440decba611fb6665099ea729c5d8b60b42b55644f2be04030b74123ceead2\": container with ID starting with f6440decba611fb6665099ea729c5d8b60b42b55644f2be04030b74123ceead2 not found: ID does not exist" containerID="f6440decba611fb6665099ea729c5d8b60b42b55644f2be04030b74123ceead2" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.174838 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6440decba611fb6665099ea729c5d8b60b42b55644f2be04030b74123ceead2"} err="failed to get container status \"f6440decba611fb6665099ea729c5d8b60b42b55644f2be04030b74123ceead2\": rpc error: code = NotFound desc = could not find container \"f6440decba611fb6665099ea729c5d8b60b42b55644f2be04030b74123ceead2\": container with ID starting with f6440decba611fb6665099ea729c5d8b60b42b55644f2be04030b74123ceead2 not found: ID does not exist" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.174917 4702 scope.go:117] "RemoveContainer" containerID="cd9ebd18fca903bcc5c3910acd17c494abdac265821f0092b48049e2d699a783" Nov 25 14:15:35 crc kubenswrapper[4702]: E1125 14:15:35.175446 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd9ebd18fca903bcc5c3910acd17c494abdac265821f0092b48049e2d699a783\": container with ID starting with cd9ebd18fca903bcc5c3910acd17c494abdac265821f0092b48049e2d699a783 not found: ID does not exist" containerID="cd9ebd18fca903bcc5c3910acd17c494abdac265821f0092b48049e2d699a783" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.175575 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd9ebd18fca903bcc5c3910acd17c494abdac265821f0092b48049e2d699a783"} err="failed to get container status \"cd9ebd18fca903bcc5c3910acd17c494abdac265821f0092b48049e2d699a783\": rpc error: code = NotFound desc = could not find container \"cd9ebd18fca903bcc5c3910acd17c494abdac265821f0092b48049e2d699a783\": container with ID starting with cd9ebd18fca903bcc5c3910acd17c494abdac265821f0092b48049e2d699a783 not found: ID does not exist" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.175643 4702 scope.go:117] "RemoveContainer" containerID="5df1dfc91583f02a0504817678e09b0d330f585b91b0df332270cca5ebc47573" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.176495 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5df1dfc91583f02a0504817678e09b0d330f585b91b0df332270cca5ebc47573"} err="failed to get container status \"5df1dfc91583f02a0504817678e09b0d330f585b91b0df332270cca5ebc47573\": rpc error: code = NotFound desc = could not find container \"5df1dfc91583f02a0504817678e09b0d330f585b91b0df332270cca5ebc47573\": container with ID starting with 5df1dfc91583f02a0504817678e09b0d330f585b91b0df332270cca5ebc47573 not found: ID does not exist" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.176539 4702 scope.go:117] "RemoveContainer" containerID="f6440decba611fb6665099ea729c5d8b60b42b55644f2be04030b74123ceead2" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.177182 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6440decba611fb6665099ea729c5d8b60b42b55644f2be04030b74123ceead2"} err="failed to get container status \"f6440decba611fb6665099ea729c5d8b60b42b55644f2be04030b74123ceead2\": rpc error: code = NotFound desc = could not find container \"f6440decba611fb6665099ea729c5d8b60b42b55644f2be04030b74123ceead2\": container with ID starting with f6440decba611fb6665099ea729c5d8b60b42b55644f2be04030b74123ceead2 not found: ID does not exist" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.177216 4702 scope.go:117] "RemoveContainer" containerID="cd9ebd18fca903bcc5c3910acd17c494abdac265821f0092b48049e2d699a783" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.177471 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd9ebd18fca903bcc5c3910acd17c494abdac265821f0092b48049e2d699a783"} err="failed to get container status \"cd9ebd18fca903bcc5c3910acd17c494abdac265821f0092b48049e2d699a783\": rpc error: code = NotFound desc = could not find container \"cd9ebd18fca903bcc5c3910acd17c494abdac265821f0092b48049e2d699a783\": container with ID starting with cd9ebd18fca903bcc5c3910acd17c494abdac265821f0092b48049e2d699a783 not found: ID does not exist" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.177495 4702 scope.go:117] "RemoveContainer" containerID="5df1dfc91583f02a0504817678e09b0d330f585b91b0df332270cca5ebc47573" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.178078 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5df1dfc91583f02a0504817678e09b0d330f585b91b0df332270cca5ebc47573"} err="failed to get container status \"5df1dfc91583f02a0504817678e09b0d330f585b91b0df332270cca5ebc47573\": rpc error: code = NotFound desc = could not find container \"5df1dfc91583f02a0504817678e09b0d330f585b91b0df332270cca5ebc47573\": container with ID starting with 5df1dfc91583f02a0504817678e09b0d330f585b91b0df332270cca5ebc47573 not found: ID does not exist" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.178107 4702 scope.go:117] "RemoveContainer" containerID="f6440decba611fb6665099ea729c5d8b60b42b55644f2be04030b74123ceead2" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.178608 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6440decba611fb6665099ea729c5d8b60b42b55644f2be04030b74123ceead2"} err="failed to get container status \"f6440decba611fb6665099ea729c5d8b60b42b55644f2be04030b74123ceead2\": rpc error: code = NotFound desc = could not find container \"f6440decba611fb6665099ea729c5d8b60b42b55644f2be04030b74123ceead2\": container with ID starting with f6440decba611fb6665099ea729c5d8b60b42b55644f2be04030b74123ceead2 not found: ID does not exist" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.178707 4702 scope.go:117] "RemoveContainer" containerID="cd9ebd18fca903bcc5c3910acd17c494abdac265821f0092b48049e2d699a783" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.179001 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd9ebd18fca903bcc5c3910acd17c494abdac265821f0092b48049e2d699a783"} err="failed to get container status \"cd9ebd18fca903bcc5c3910acd17c494abdac265821f0092b48049e2d699a783\": rpc error: code = NotFound desc = could not find container \"cd9ebd18fca903bcc5c3910acd17c494abdac265821f0092b48049e2d699a783\": container with ID starting with cd9ebd18fca903bcc5c3910acd17c494abdac265821f0092b48049e2d699a783 not found: ID does not exist" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.219337 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ba715ae-64e0-4810-b14d-f0c73b02f65d-config-data" (OuterVolumeSpecName: "config-data") pod "2ba715ae-64e0-4810-b14d-f0c73b02f65d" (UID: "2ba715ae-64e0-4810-b14d-f0c73b02f65d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.221784 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-config-data" (OuterVolumeSpecName: "config-data") pod "50e365be-ffbe-4bf9-8eec-19cca5ca20cb" (UID: "50e365be-ffbe-4bf9-8eec-19cca5ca20cb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.240896 4702 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ba715ae-64e0-4810-b14d-f0c73b02f65d-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.240957 4702 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.240972 4702 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.240983 4702 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.240993 4702 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.241012 4702 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.241022 4702 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-sys\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.241032 4702 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ba715ae-64e0-4810-b14d-f0c73b02f65d-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.241043 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vcd6s\" (UniqueName: \"kubernetes.io/projected/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-kube-api-access-vcd6s\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.241054 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbxxm\" (UniqueName: \"kubernetes.io/projected/2ba715ae-64e0-4810-b14d-f0c73b02f65d-kube-api-access-rbxxm\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.241070 4702 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.241084 4702 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.241097 4702 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50e365be-ffbe-4bf9-8eec-19cca5ca20cb-logs\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.254403 4702 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.256553 4702 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.258278 4702 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.258771 4702 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.342363 4702 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.342396 4702 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.342405 4702 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.342414 4702 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.378692 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.387500 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.394597 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.402767 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.810638 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b5b9839-7217-4d14-a6e3-c8e62579f40e" path="/var/lib/kubelet/pods/2b5b9839-7217-4d14-a6e3-c8e62579f40e/volumes" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.811816 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ba715ae-64e0-4810-b14d-f0c73b02f65d" path="/var/lib/kubelet/pods/2ba715ae-64e0-4810-b14d-f0c73b02f65d/volumes" Nov 25 14:15:35 crc kubenswrapper[4702]: I1125 14:15:35.812983 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50e365be-ffbe-4bf9-8eec-19cca5ca20cb" path="/var/lib/kubelet/pods/50e365be-ffbe-4bf9-8eec-19cca5ca20cb/volumes" Nov 25 14:15:36 crc kubenswrapper[4702]: I1125 14:15:36.042890 4702 generic.go:334] "Generic (PLEG): container finished" podID="e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4" containerID="0536a3dac6b639535a8e04cbcbdb471a210f9dd77cf0b1d4391a71227e1d1e17" exitCode=0 Nov 25 14:15:36 crc kubenswrapper[4702]: I1125 14:15:36.042938 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance9a19-account-delete-xggfm" event={"ID":"e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4","Type":"ContainerDied","Data":"0536a3dac6b639535a8e04cbcbdb471a210f9dd77cf0b1d4391a71227e1d1e17"} Nov 25 14:15:36 crc kubenswrapper[4702]: I1125 14:15:36.801364 4702 scope.go:117] "RemoveContainer" containerID="32273717263d7fcc99e8193a439dd976da66869c55bb4633d9314cacb6ab92a1" Nov 25 14:15:36 crc kubenswrapper[4702]: E1125 14:15:36.801699 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qqr76_openshift-machine-config-operator(3c491818-b0bb-4d82-a031-96e2dfac8c27)\"" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" Nov 25 14:15:37 crc kubenswrapper[4702]: I1125 14:15:37.373876 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance9a19-account-delete-xggfm" Nov 25 14:15:37 crc kubenswrapper[4702]: I1125 14:15:37.487185 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4-operator-scripts\") pod \"e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4\" (UID: \"e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4\") " Nov 25 14:15:37 crc kubenswrapper[4702]: I1125 14:15:37.487471 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hsdj9\" (UniqueName: \"kubernetes.io/projected/e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4-kube-api-access-hsdj9\") pod \"e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4\" (UID: \"e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4\") " Nov 25 14:15:37 crc kubenswrapper[4702]: I1125 14:15:37.488518 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4" (UID: "e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 14:15:37 crc kubenswrapper[4702]: I1125 14:15:37.494607 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4-kube-api-access-hsdj9" (OuterVolumeSpecName: "kube-api-access-hsdj9") pod "e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4" (UID: "e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4"). InnerVolumeSpecName "kube-api-access-hsdj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:15:37 crc kubenswrapper[4702]: I1125 14:15:37.588814 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hsdj9\" (UniqueName: \"kubernetes.io/projected/e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4-kube-api-access-hsdj9\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:37 crc kubenswrapper[4702]: I1125 14:15:37.588853 4702 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.061821 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance9a19-account-delete-xggfm" event={"ID":"e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4","Type":"ContainerDied","Data":"a8a04f149a6f55bf9228ad076b5fd5dcf76e0299f393d6b3fba6ee6f38524ca7"} Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.061876 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a8a04f149a6f55bf9228ad076b5fd5dcf76e0299f393d6b3fba6ee6f38524ca7" Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.061876 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance9a19-account-delete-xggfm" Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.252100 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5fsgg"] Nov 25 14:15:38 crc kubenswrapper[4702]: E1125 14:15:38.253668 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50e365be-ffbe-4bf9-8eec-19cca5ca20cb" containerName="glance-httpd" Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.253697 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="50e365be-ffbe-4bf9-8eec-19cca5ca20cb" containerName="glance-httpd" Nov 25 14:15:38 crc kubenswrapper[4702]: E1125 14:15:38.253707 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4" containerName="mariadb-account-delete" Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.253909 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4" containerName="mariadb-account-delete" Nov 25 14:15:38 crc kubenswrapper[4702]: E1125 14:15:38.253918 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ba715ae-64e0-4810-b14d-f0c73b02f65d" containerName="glance-httpd" Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.253923 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ba715ae-64e0-4810-b14d-f0c73b02f65d" containerName="glance-httpd" Nov 25 14:15:38 crc kubenswrapper[4702]: E1125 14:15:38.253936 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50e365be-ffbe-4bf9-8eec-19cca5ca20cb" containerName="glance-api" Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.253942 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="50e365be-ffbe-4bf9-8eec-19cca5ca20cb" containerName="glance-api" Nov 25 14:15:38 crc kubenswrapper[4702]: E1125 14:15:38.253956 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ba715ae-64e0-4810-b14d-f0c73b02f65d" containerName="glance-log" Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.253962 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ba715ae-64e0-4810-b14d-f0c73b02f65d" containerName="glance-log" Nov 25 14:15:38 crc kubenswrapper[4702]: E1125 14:15:38.253973 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50e365be-ffbe-4bf9-8eec-19cca5ca20cb" containerName="glance-log" Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.253978 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="50e365be-ffbe-4bf9-8eec-19cca5ca20cb" containerName="glance-log" Nov 25 14:15:38 crc kubenswrapper[4702]: E1125 14:15:38.253992 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ba715ae-64e0-4810-b14d-f0c73b02f65d" containerName="glance-api" Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.253997 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ba715ae-64e0-4810-b14d-f0c73b02f65d" containerName="glance-api" Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.254109 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4" containerName="mariadb-account-delete" Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.254118 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="50e365be-ffbe-4bf9-8eec-19cca5ca20cb" containerName="glance-log" Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.254126 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ba715ae-64e0-4810-b14d-f0c73b02f65d" containerName="glance-httpd" Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.254137 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="50e365be-ffbe-4bf9-8eec-19cca5ca20cb" containerName="glance-httpd" Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.254143 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="50e365be-ffbe-4bf9-8eec-19cca5ca20cb" containerName="glance-api" Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.254155 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ba715ae-64e0-4810-b14d-f0c73b02f65d" containerName="glance-log" Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.254163 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ba715ae-64e0-4810-b14d-f0c73b02f65d" containerName="glance-api" Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.255294 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5fsgg" Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.271490 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5fsgg"] Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.303038 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4-utilities\") pod \"redhat-operators-5fsgg\" (UID: \"5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4\") " pod="openshift-marketplace/redhat-operators-5fsgg" Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.303099 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4-catalog-content\") pod \"redhat-operators-5fsgg\" (UID: \"5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4\") " pod="openshift-marketplace/redhat-operators-5fsgg" Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.303118 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lg27\" (UniqueName: \"kubernetes.io/projected/5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4-kube-api-access-5lg27\") pod \"redhat-operators-5fsgg\" (UID: \"5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4\") " pod="openshift-marketplace/redhat-operators-5fsgg" Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.404950 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4-utilities\") pod \"redhat-operators-5fsgg\" (UID: \"5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4\") " pod="openshift-marketplace/redhat-operators-5fsgg" Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.405000 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4-catalog-content\") pod \"redhat-operators-5fsgg\" (UID: \"5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4\") " pod="openshift-marketplace/redhat-operators-5fsgg" Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.405024 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lg27\" (UniqueName: \"kubernetes.io/projected/5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4-kube-api-access-5lg27\") pod \"redhat-operators-5fsgg\" (UID: \"5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4\") " pod="openshift-marketplace/redhat-operators-5fsgg" Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.405854 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4-utilities\") pod \"redhat-operators-5fsgg\" (UID: \"5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4\") " pod="openshift-marketplace/redhat-operators-5fsgg" Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.405893 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4-catalog-content\") pod \"redhat-operators-5fsgg\" (UID: \"5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4\") " pod="openshift-marketplace/redhat-operators-5fsgg" Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.422355 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lg27\" (UniqueName: \"kubernetes.io/projected/5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4-kube-api-access-5lg27\") pod \"redhat-operators-5fsgg\" (UID: \"5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4\") " pod="openshift-marketplace/redhat-operators-5fsgg" Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.575993 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5fsgg" Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.869293 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-zbqwl"] Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.874558 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-zbqwl"] Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.885730 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-9a19-account-create-update-g98xf"] Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.891757 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance9a19-account-delete-xggfm"] Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.897898 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-9a19-account-create-update-g98xf"] Nov 25 14:15:38 crc kubenswrapper[4702]: I1125 14:15:38.904166 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance9a19-account-delete-xggfm"] Nov 25 14:15:39 crc kubenswrapper[4702]: I1125 14:15:39.006439 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5fsgg"] Nov 25 14:15:39 crc kubenswrapper[4702]: W1125 14:15:39.011397 4702 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5aa65c5d_3c6c_4fe8_9fff_ca951d8044e4.slice/crio-6fe7954c2b58c33a47450e5f0e3ed7188ed5a477ca5ff08ecbc505a158fdfb37 WatchSource:0}: Error finding container 6fe7954c2b58c33a47450e5f0e3ed7188ed5a477ca5ff08ecbc505a158fdfb37: Status 404 returned error can't find the container with id 6fe7954c2b58c33a47450e5f0e3ed7188ed5a477ca5ff08ecbc505a158fdfb37 Nov 25 14:15:39 crc kubenswrapper[4702]: I1125 14:15:39.073143 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5fsgg" event={"ID":"5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4","Type":"ContainerStarted","Data":"6fe7954c2b58c33a47450e5f0e3ed7188ed5a477ca5ff08ecbc505a158fdfb37"} Nov 25 14:15:40 crc kubenswrapper[4702]: I1125 14:15:40.049298 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a261226b-a600-4a09-a99f-1d886ddf8e6f" path="/var/lib/kubelet/pods/a261226b-a600-4a09-a99f-1d886ddf8e6f/volumes" Nov 25 14:15:40 crc kubenswrapper[4702]: I1125 14:15:40.084502 4702 generic.go:334] "Generic (PLEG): container finished" podID="5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4" containerID="866f8825cb6d7155d8b7263e0471723998955da731f365b5692f829dfb64d15c" exitCode=0 Nov 25 14:15:40 crc kubenswrapper[4702]: I1125 14:15:40.120691 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4" path="/var/lib/kubelet/pods/e7c3ae73-4ad8-4e64-a472-b60fc4f20ab4/volumes" Nov 25 14:15:40 crc kubenswrapper[4702]: I1125 14:15:40.121410 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe67604e-401d-4050-8c57-ed41d56b72f0" path="/var/lib/kubelet/pods/fe67604e-401d-4050-8c57-ed41d56b72f0/volumes" Nov 25 14:15:40 crc kubenswrapper[4702]: I1125 14:15:40.121887 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5fsgg" event={"ID":"5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4","Type":"ContainerDied","Data":"866f8825cb6d7155d8b7263e0471723998955da731f365b5692f829dfb64d15c"} Nov 25 14:15:43 crc kubenswrapper[4702]: I1125 14:15:43.106726 4702 generic.go:334] "Generic (PLEG): container finished" podID="5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4" containerID="172b9e2959eaae23d76ca051d3d0f5966c0a025d2642a9ee1f44f91d9f31f512" exitCode=0 Nov 25 14:15:43 crc kubenswrapper[4702]: I1125 14:15:43.106783 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5fsgg" event={"ID":"5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4","Type":"ContainerDied","Data":"172b9e2959eaae23d76ca051d3d0f5966c0a025d2642a9ee1f44f91d9f31f512"} Nov 25 14:15:44 crc kubenswrapper[4702]: I1125 14:15:44.121333 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5fsgg" event={"ID":"5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4","Type":"ContainerStarted","Data":"66488d3e4c9f06d4cf706d1225801471cb82a15738cc08b6b9e60bffec5c94c7"} Nov 25 14:15:45 crc kubenswrapper[4702]: I1125 14:15:45.150533 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5fsgg" podStartSLOduration=3.315849396 podStartE2EDuration="7.150509902s" podCreationTimestamp="2025-11-25 14:15:38 +0000 UTC" firstStartedPulling="2025-11-25 14:15:40.08603903 +0000 UTC m=+1749.248689480" lastFinishedPulling="2025-11-25 14:15:43.920699526 +0000 UTC m=+1753.083349986" observedRunningTime="2025-11-25 14:15:45.144138975 +0000 UTC m=+1754.306789425" watchObservedRunningTime="2025-11-25 14:15:45.150509902 +0000 UTC m=+1754.313160352" Nov 25 14:15:47 crc kubenswrapper[4702]: I1125 14:15:47.801904 4702 scope.go:117] "RemoveContainer" containerID="32273717263d7fcc99e8193a439dd976da66869c55bb4633d9314cacb6ab92a1" Nov 25 14:15:47 crc kubenswrapper[4702]: E1125 14:15:47.802394 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qqr76_openshift-machine-config-operator(3c491818-b0bb-4d82-a031-96e2dfac8c27)\"" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" Nov 25 14:15:48 crc kubenswrapper[4702]: I1125 14:15:48.576128 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5fsgg" Nov 25 14:15:48 crc kubenswrapper[4702]: I1125 14:15:48.576487 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5fsgg" Nov 25 14:15:48 crc kubenswrapper[4702]: I1125 14:15:48.621572 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5fsgg" Nov 25 14:15:49 crc kubenswrapper[4702]: I1125 14:15:49.202805 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5fsgg" Nov 25 14:15:50 crc kubenswrapper[4702]: I1125 14:15:50.842181 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5fsgg"] Nov 25 14:15:51 crc kubenswrapper[4702]: I1125 14:15:51.175950 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5fsgg" podUID="5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4" containerName="registry-server" containerID="cri-o://66488d3e4c9f06d4cf706d1225801471cb82a15738cc08b6b9e60bffec5c94c7" gracePeriod=2 Nov 25 14:15:52 crc kubenswrapper[4702]: I1125 14:15:52.901907 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5fsgg" Nov 25 14:15:53 crc kubenswrapper[4702]: I1125 14:15:53.016219 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4-catalog-content\") pod \"5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4\" (UID: \"5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4\") " Nov 25 14:15:53 crc kubenswrapper[4702]: I1125 14:15:53.016317 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lg27\" (UniqueName: \"kubernetes.io/projected/5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4-kube-api-access-5lg27\") pod \"5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4\" (UID: \"5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4\") " Nov 25 14:15:53 crc kubenswrapper[4702]: I1125 14:15:53.016347 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4-utilities\") pod \"5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4\" (UID: \"5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4\") " Nov 25 14:15:53 crc kubenswrapper[4702]: I1125 14:15:53.017631 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4-utilities" (OuterVolumeSpecName: "utilities") pod "5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4" (UID: "5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:15:53 crc kubenswrapper[4702]: I1125 14:15:53.021371 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4-kube-api-access-5lg27" (OuterVolumeSpecName: "kube-api-access-5lg27") pod "5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4" (UID: "5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4"). InnerVolumeSpecName "kube-api-access-5lg27". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:15:53 crc kubenswrapper[4702]: I1125 14:15:53.112484 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4" (UID: "5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:15:53 crc kubenswrapper[4702]: I1125 14:15:53.118605 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:53 crc kubenswrapper[4702]: I1125 14:15:53.118650 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lg27\" (UniqueName: \"kubernetes.io/projected/5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4-kube-api-access-5lg27\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:53 crc kubenswrapper[4702]: I1125 14:15:53.118664 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 14:15:53 crc kubenswrapper[4702]: I1125 14:15:53.192430 4702 generic.go:334] "Generic (PLEG): container finished" podID="5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4" containerID="66488d3e4c9f06d4cf706d1225801471cb82a15738cc08b6b9e60bffec5c94c7" exitCode=0 Nov 25 14:15:53 crc kubenswrapper[4702]: I1125 14:15:53.192485 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5fsgg" event={"ID":"5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4","Type":"ContainerDied","Data":"66488d3e4c9f06d4cf706d1225801471cb82a15738cc08b6b9e60bffec5c94c7"} Nov 25 14:15:53 crc kubenswrapper[4702]: I1125 14:15:53.192534 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5fsgg" event={"ID":"5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4","Type":"ContainerDied","Data":"6fe7954c2b58c33a47450e5f0e3ed7188ed5a477ca5ff08ecbc505a158fdfb37"} Nov 25 14:15:53 crc kubenswrapper[4702]: I1125 14:15:53.192557 4702 scope.go:117] "RemoveContainer" containerID="66488d3e4c9f06d4cf706d1225801471cb82a15738cc08b6b9e60bffec5c94c7" Nov 25 14:15:53 crc kubenswrapper[4702]: I1125 14:15:53.192592 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5fsgg" Nov 25 14:15:53 crc kubenswrapper[4702]: I1125 14:15:53.215146 4702 scope.go:117] "RemoveContainer" containerID="172b9e2959eaae23d76ca051d3d0f5966c0a025d2642a9ee1f44f91d9f31f512" Nov 25 14:15:53 crc kubenswrapper[4702]: I1125 14:15:53.233195 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5fsgg"] Nov 25 14:15:53 crc kubenswrapper[4702]: I1125 14:15:53.240867 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5fsgg"] Nov 25 14:15:53 crc kubenswrapper[4702]: I1125 14:15:53.256367 4702 scope.go:117] "RemoveContainer" containerID="866f8825cb6d7155d8b7263e0471723998955da731f365b5692f829dfb64d15c" Nov 25 14:15:53 crc kubenswrapper[4702]: I1125 14:15:53.276644 4702 scope.go:117] "RemoveContainer" containerID="66488d3e4c9f06d4cf706d1225801471cb82a15738cc08b6b9e60bffec5c94c7" Nov 25 14:15:53 crc kubenswrapper[4702]: E1125 14:15:53.277309 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66488d3e4c9f06d4cf706d1225801471cb82a15738cc08b6b9e60bffec5c94c7\": container with ID starting with 66488d3e4c9f06d4cf706d1225801471cb82a15738cc08b6b9e60bffec5c94c7 not found: ID does not exist" containerID="66488d3e4c9f06d4cf706d1225801471cb82a15738cc08b6b9e60bffec5c94c7" Nov 25 14:15:53 crc kubenswrapper[4702]: I1125 14:15:53.277372 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66488d3e4c9f06d4cf706d1225801471cb82a15738cc08b6b9e60bffec5c94c7"} err="failed to get container status \"66488d3e4c9f06d4cf706d1225801471cb82a15738cc08b6b9e60bffec5c94c7\": rpc error: code = NotFound desc = could not find container \"66488d3e4c9f06d4cf706d1225801471cb82a15738cc08b6b9e60bffec5c94c7\": container with ID starting with 66488d3e4c9f06d4cf706d1225801471cb82a15738cc08b6b9e60bffec5c94c7 not found: ID does not exist" Nov 25 14:15:53 crc kubenswrapper[4702]: I1125 14:15:53.277410 4702 scope.go:117] "RemoveContainer" containerID="172b9e2959eaae23d76ca051d3d0f5966c0a025d2642a9ee1f44f91d9f31f512" Nov 25 14:15:53 crc kubenswrapper[4702]: E1125 14:15:53.277923 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"172b9e2959eaae23d76ca051d3d0f5966c0a025d2642a9ee1f44f91d9f31f512\": container with ID starting with 172b9e2959eaae23d76ca051d3d0f5966c0a025d2642a9ee1f44f91d9f31f512 not found: ID does not exist" containerID="172b9e2959eaae23d76ca051d3d0f5966c0a025d2642a9ee1f44f91d9f31f512" Nov 25 14:15:53 crc kubenswrapper[4702]: I1125 14:15:53.277957 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"172b9e2959eaae23d76ca051d3d0f5966c0a025d2642a9ee1f44f91d9f31f512"} err="failed to get container status \"172b9e2959eaae23d76ca051d3d0f5966c0a025d2642a9ee1f44f91d9f31f512\": rpc error: code = NotFound desc = could not find container \"172b9e2959eaae23d76ca051d3d0f5966c0a025d2642a9ee1f44f91d9f31f512\": container with ID starting with 172b9e2959eaae23d76ca051d3d0f5966c0a025d2642a9ee1f44f91d9f31f512 not found: ID does not exist" Nov 25 14:15:53 crc kubenswrapper[4702]: I1125 14:15:53.277975 4702 scope.go:117] "RemoveContainer" containerID="866f8825cb6d7155d8b7263e0471723998955da731f365b5692f829dfb64d15c" Nov 25 14:15:53 crc kubenswrapper[4702]: E1125 14:15:53.278293 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"866f8825cb6d7155d8b7263e0471723998955da731f365b5692f829dfb64d15c\": container with ID starting with 866f8825cb6d7155d8b7263e0471723998955da731f365b5692f829dfb64d15c not found: ID does not exist" containerID="866f8825cb6d7155d8b7263e0471723998955da731f365b5692f829dfb64d15c" Nov 25 14:15:53 crc kubenswrapper[4702]: I1125 14:15:53.278333 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"866f8825cb6d7155d8b7263e0471723998955da731f365b5692f829dfb64d15c"} err="failed to get container status \"866f8825cb6d7155d8b7263e0471723998955da731f365b5692f829dfb64d15c\": rpc error: code = NotFound desc = could not find container \"866f8825cb6d7155d8b7263e0471723998955da731f365b5692f829dfb64d15c\": container with ID starting with 866f8825cb6d7155d8b7263e0471723998955da731f365b5692f829dfb64d15c not found: ID does not exist" Nov 25 14:15:53 crc kubenswrapper[4702]: I1125 14:15:53.809812 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4" path="/var/lib/kubelet/pods/5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4/volumes" Nov 25 14:15:59 crc kubenswrapper[4702]: I1125 14:15:59.801728 4702 scope.go:117] "RemoveContainer" containerID="32273717263d7fcc99e8193a439dd976da66869c55bb4633d9314cacb6ab92a1" Nov 25 14:15:59 crc kubenswrapper[4702]: E1125 14:15:59.802194 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qqr76_openshift-machine-config-operator(3c491818-b0bb-4d82-a031-96e2dfac8c27)\"" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" Nov 25 14:16:04 crc kubenswrapper[4702]: I1125 14:16:04.953661 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="2ba715ae-64e0-4810-b14d-f0c73b02f65d" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.0.113:9292/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 14:16:04 crc kubenswrapper[4702]: I1125 14:16:04.955992 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="2ba715ae-64e0-4810-b14d-f0c73b02f65d" containerName="glance-api" probeResult="failure" output="Get \"http://10.217.0.113:9292/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 14:16:04 crc kubenswrapper[4702]: I1125 14:16:04.955982 4702 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="2ba715ae-64e0-4810-b14d-f0c73b02f65d" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.0.113:9292/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 14:16:12 crc kubenswrapper[4702]: I1125 14:16:12.801678 4702 scope.go:117] "RemoveContainer" containerID="32273717263d7fcc99e8193a439dd976da66869c55bb4633d9314cacb6ab92a1" Nov 25 14:16:12 crc kubenswrapper[4702]: E1125 14:16:12.802976 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qqr76_openshift-machine-config-operator(3c491818-b0bb-4d82-a031-96e2dfac8c27)\"" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" Nov 25 14:16:12 crc kubenswrapper[4702]: I1125 14:16:12.889344 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-k9pzw/must-gather-wsn6h"] Nov 25 14:16:12 crc kubenswrapper[4702]: E1125 14:16:12.890882 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4" containerName="extract-utilities" Nov 25 14:16:12 crc kubenswrapper[4702]: I1125 14:16:12.890966 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4" containerName="extract-utilities" Nov 25 14:16:12 crc kubenswrapper[4702]: E1125 14:16:12.891018 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4" containerName="extract-content" Nov 25 14:16:12 crc kubenswrapper[4702]: I1125 14:16:12.891032 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4" containerName="extract-content" Nov 25 14:16:12 crc kubenswrapper[4702]: E1125 14:16:12.891068 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4" containerName="registry-server" Nov 25 14:16:12 crc kubenswrapper[4702]: I1125 14:16:12.891080 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4" containerName="registry-server" Nov 25 14:16:12 crc kubenswrapper[4702]: I1125 14:16:12.892550 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="5aa65c5d-3c6c-4fe8-9fff-ca951d8044e4" containerName="registry-server" Nov 25 14:16:12 crc kubenswrapper[4702]: I1125 14:16:12.901603 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k9pzw/must-gather-wsn6h" Nov 25 14:16:12 crc kubenswrapper[4702]: I1125 14:16:12.908676 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-k9pzw"/"kube-root-ca.crt" Nov 25 14:16:12 crc kubenswrapper[4702]: I1125 14:16:12.908735 4702 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-k9pzw"/"default-dockercfg-rxdh4" Nov 25 14:16:12 crc kubenswrapper[4702]: I1125 14:16:12.910882 4702 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-k9pzw"/"openshift-service-ca.crt" Nov 25 14:16:12 crc kubenswrapper[4702]: I1125 14:16:12.961303 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-k9pzw/must-gather-wsn6h"] Nov 25 14:16:13 crc kubenswrapper[4702]: I1125 14:16:13.038038 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5bd1aa0b-893d-4550-85eb-3ee257c7b355-must-gather-output\") pod \"must-gather-wsn6h\" (UID: \"5bd1aa0b-893d-4550-85eb-3ee257c7b355\") " pod="openshift-must-gather-k9pzw/must-gather-wsn6h" Nov 25 14:16:13 crc kubenswrapper[4702]: I1125 14:16:13.038158 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hksw6\" (UniqueName: \"kubernetes.io/projected/5bd1aa0b-893d-4550-85eb-3ee257c7b355-kube-api-access-hksw6\") pod \"must-gather-wsn6h\" (UID: \"5bd1aa0b-893d-4550-85eb-3ee257c7b355\") " pod="openshift-must-gather-k9pzw/must-gather-wsn6h" Nov 25 14:16:13 crc kubenswrapper[4702]: I1125 14:16:13.139678 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5bd1aa0b-893d-4550-85eb-3ee257c7b355-must-gather-output\") pod \"must-gather-wsn6h\" (UID: \"5bd1aa0b-893d-4550-85eb-3ee257c7b355\") " pod="openshift-must-gather-k9pzw/must-gather-wsn6h" Nov 25 14:16:13 crc kubenswrapper[4702]: I1125 14:16:13.139802 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hksw6\" (UniqueName: \"kubernetes.io/projected/5bd1aa0b-893d-4550-85eb-3ee257c7b355-kube-api-access-hksw6\") pod \"must-gather-wsn6h\" (UID: \"5bd1aa0b-893d-4550-85eb-3ee257c7b355\") " pod="openshift-must-gather-k9pzw/must-gather-wsn6h" Nov 25 14:16:13 crc kubenswrapper[4702]: I1125 14:16:13.140409 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5bd1aa0b-893d-4550-85eb-3ee257c7b355-must-gather-output\") pod \"must-gather-wsn6h\" (UID: \"5bd1aa0b-893d-4550-85eb-3ee257c7b355\") " pod="openshift-must-gather-k9pzw/must-gather-wsn6h" Nov 25 14:16:13 crc kubenswrapper[4702]: I1125 14:16:13.163172 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hksw6\" (UniqueName: \"kubernetes.io/projected/5bd1aa0b-893d-4550-85eb-3ee257c7b355-kube-api-access-hksw6\") pod \"must-gather-wsn6h\" (UID: \"5bd1aa0b-893d-4550-85eb-3ee257c7b355\") " pod="openshift-must-gather-k9pzw/must-gather-wsn6h" Nov 25 14:16:13 crc kubenswrapper[4702]: I1125 14:16:13.240727 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k9pzw/must-gather-wsn6h" Nov 25 14:16:13 crc kubenswrapper[4702]: I1125 14:16:13.732917 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-k9pzw/must-gather-wsn6h"] Nov 25 14:16:13 crc kubenswrapper[4702]: I1125 14:16:13.761508 4702 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 14:16:14 crc kubenswrapper[4702]: I1125 14:16:14.356670 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-k9pzw/must-gather-wsn6h" event={"ID":"5bd1aa0b-893d-4550-85eb-3ee257c7b355","Type":"ContainerStarted","Data":"ca3a19068192db6cbdc519cda282633d26fef2399bc1849ffa91f943101653f6"} Nov 25 14:16:18 crc kubenswrapper[4702]: I1125 14:16:18.390084 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-k9pzw/must-gather-wsn6h" event={"ID":"5bd1aa0b-893d-4550-85eb-3ee257c7b355","Type":"ContainerStarted","Data":"343a84620bdc6ee504942e6a904d4bf9437e49094c13f6176cabb793a0524bbb"} Nov 25 14:16:18 crc kubenswrapper[4702]: I1125 14:16:18.390659 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-k9pzw/must-gather-wsn6h" event={"ID":"5bd1aa0b-893d-4550-85eb-3ee257c7b355","Type":"ContainerStarted","Data":"d89d654239cd9f7f93a9910cdfea090141ef0f2cb7223575974745f290431aa1"} Nov 25 14:16:18 crc kubenswrapper[4702]: I1125 14:16:18.412198 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-k9pzw/must-gather-wsn6h" podStartSLOduration=2.529977759 podStartE2EDuration="6.41218044s" podCreationTimestamp="2025-11-25 14:16:12 +0000 UTC" firstStartedPulling="2025-11-25 14:16:13.761130883 +0000 UTC m=+1782.923781333" lastFinishedPulling="2025-11-25 14:16:17.643333564 +0000 UTC m=+1786.805984014" observedRunningTime="2025-11-25 14:16:18.408500879 +0000 UTC m=+1787.571151349" watchObservedRunningTime="2025-11-25 14:16:18.41218044 +0000 UTC m=+1787.574830890" Nov 25 14:16:26 crc kubenswrapper[4702]: I1125 14:16:26.800945 4702 scope.go:117] "RemoveContainer" containerID="32273717263d7fcc99e8193a439dd976da66869c55bb4633d9314cacb6ab92a1" Nov 25 14:16:26 crc kubenswrapper[4702]: E1125 14:16:26.801673 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qqr76_openshift-machine-config-operator(3c491818-b0bb-4d82-a031-96e2dfac8c27)\"" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" Nov 25 14:16:33 crc kubenswrapper[4702]: I1125 14:16:33.681567 4702 scope.go:117] "RemoveContainer" containerID="4c7e021df80ed355ad656829dbe1d50152175b26e8f73f74f792b7105673d42f" Nov 25 14:16:33 crc kubenswrapper[4702]: I1125 14:16:33.736223 4702 scope.go:117] "RemoveContainer" containerID="b9e3730cdee2ae8bac9ed922e1d63339550bc037a43007735d07fcd58c75e95e" Nov 25 14:16:33 crc kubenswrapper[4702]: I1125 14:16:33.755960 4702 scope.go:117] "RemoveContainer" containerID="ce06422e8fcd751bfcda6a78556c07b9613857458992640ca90dce8f564cdb7d" Nov 25 14:16:39 crc kubenswrapper[4702]: I1125 14:16:39.801227 4702 scope.go:117] "RemoveContainer" containerID="32273717263d7fcc99e8193a439dd976da66869c55bb4633d9314cacb6ab92a1" Nov 25 14:16:39 crc kubenswrapper[4702]: E1125 14:16:39.803074 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qqr76_openshift-machine-config-operator(3c491818-b0bb-4d82-a031-96e2dfac8c27)\"" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" Nov 25 14:16:50 crc kubenswrapper[4702]: I1125 14:16:50.800900 4702 scope.go:117] "RemoveContainer" containerID="32273717263d7fcc99e8193a439dd976da66869c55bb4633d9314cacb6ab92a1" Nov 25 14:16:50 crc kubenswrapper[4702]: E1125 14:16:50.801613 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qqr76_openshift-machine-config-operator(3c491818-b0bb-4d82-a031-96e2dfac8c27)\"" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" Nov 25 14:16:52 crc kubenswrapper[4702]: I1125 14:16:52.740614 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g_cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba/util/0.log" Nov 25 14:16:52 crc kubenswrapper[4702]: I1125 14:16:52.917337 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g_cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba/pull/0.log" Nov 25 14:16:52 crc kubenswrapper[4702]: I1125 14:16:52.919043 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g_cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba/util/0.log" Nov 25 14:16:52 crc kubenswrapper[4702]: I1125 14:16:52.920686 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g_cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba/pull/0.log" Nov 25 14:16:53 crc kubenswrapper[4702]: I1125 14:16:53.105356 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g_cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba/extract/0.log" Nov 25 14:16:53 crc kubenswrapper[4702]: I1125 14:16:53.138314 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g_cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba/pull/0.log" Nov 25 14:16:53 crc kubenswrapper[4702]: I1125 14:16:53.149392 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d9j79g_cce8eeb8-1b65-4ef8-b8f7-74ce6a1810ba/util/0.log" Nov 25 14:16:53 crc kubenswrapper[4702]: I1125 14:16:53.288955 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd_0f4314a6-9010-4789-816c-3bc52836c02e/util/0.log" Nov 25 14:16:53 crc kubenswrapper[4702]: I1125 14:16:53.470899 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd_0f4314a6-9010-4789-816c-3bc52836c02e/util/0.log" Nov 25 14:16:53 crc kubenswrapper[4702]: I1125 14:16:53.504336 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd_0f4314a6-9010-4789-816c-3bc52836c02e/pull/0.log" Nov 25 14:16:53 crc kubenswrapper[4702]: I1125 14:16:53.522006 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd_0f4314a6-9010-4789-816c-3bc52836c02e/pull/0.log" Nov 25 14:16:53 crc kubenswrapper[4702]: I1125 14:16:53.642874 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd_0f4314a6-9010-4789-816c-3bc52836c02e/util/0.log" Nov 25 14:16:53 crc kubenswrapper[4702]: I1125 14:16:53.667276 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd_0f4314a6-9010-4789-816c-3bc52836c02e/extract/0.log" Nov 25 14:16:53 crc kubenswrapper[4702]: I1125 14:16:53.727567 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f979943479rjd_0f4314a6-9010-4789-816c-3bc52836c02e/pull/0.log" Nov 25 14:16:53 crc kubenswrapper[4702]: I1125 14:16:53.837364 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9_401b6536-9504-448e-8acb-cb09c0cb0650/util/0.log" Nov 25 14:16:53 crc kubenswrapper[4702]: I1125 14:16:53.973297 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9_401b6536-9504-448e-8acb-cb09c0cb0650/pull/0.log" Nov 25 14:16:54 crc kubenswrapper[4702]: I1125 14:16:54.012143 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9_401b6536-9504-448e-8acb-cb09c0cb0650/pull/0.log" Nov 25 14:16:54 crc kubenswrapper[4702]: I1125 14:16:54.054447 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9_401b6536-9504-448e-8acb-cb09c0cb0650/util/0.log" Nov 25 14:16:54 crc kubenswrapper[4702]: I1125 14:16:54.153960 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9_401b6536-9504-448e-8acb-cb09c0cb0650/pull/0.log" Nov 25 14:16:54 crc kubenswrapper[4702]: I1125 14:16:54.183960 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9_401b6536-9504-448e-8acb-cb09c0cb0650/util/0.log" Nov 25 14:16:54 crc kubenswrapper[4702]: I1125 14:16:54.220355 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bjv7j9_401b6536-9504-448e-8acb-cb09c0cb0650/extract/0.log" Nov 25 14:16:54 crc kubenswrapper[4702]: I1125 14:16:54.341975 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5_aeb59e3f-26d5-426a-a371-cdc4efd75add/util/0.log" Nov 25 14:16:54 crc kubenswrapper[4702]: I1125 14:16:54.467957 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5_aeb59e3f-26d5-426a-a371-cdc4efd75add/util/0.log" Nov 25 14:16:54 crc kubenswrapper[4702]: I1125 14:16:54.497400 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5_aeb59e3f-26d5-426a-a371-cdc4efd75add/pull/0.log" Nov 25 14:16:54 crc kubenswrapper[4702]: I1125 14:16:54.500283 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5_aeb59e3f-26d5-426a-a371-cdc4efd75add/pull/0.log" Nov 25 14:16:54 crc kubenswrapper[4702]: I1125 14:16:54.661584 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5_aeb59e3f-26d5-426a-a371-cdc4efd75add/extract/0.log" Nov 25 14:16:54 crc kubenswrapper[4702]: I1125 14:16:54.800533 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5_aeb59e3f-26d5-426a-a371-cdc4efd75add/pull/0.log" Nov 25 14:16:54 crc kubenswrapper[4702]: I1125 14:16:54.832622 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590pvbs5_aeb59e3f-26d5-426a-a371-cdc4efd75add/util/0.log" Nov 25 14:16:54 crc kubenswrapper[4702]: I1125 14:16:54.858539 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp_015eb990-26fa-467f-b645-fd3290426356/util/0.log" Nov 25 14:16:55 crc kubenswrapper[4702]: I1125 14:16:55.023781 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp_015eb990-26fa-467f-b645-fd3290426356/pull/0.log" Nov 25 14:16:55 crc kubenswrapper[4702]: I1125 14:16:55.048293 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp_015eb990-26fa-467f-b645-fd3290426356/pull/0.log" Nov 25 14:16:55 crc kubenswrapper[4702]: I1125 14:16:55.095219 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp_015eb990-26fa-467f-b645-fd3290426356/util/0.log" Nov 25 14:16:55 crc kubenswrapper[4702]: I1125 14:16:55.324187 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp_015eb990-26fa-467f-b645-fd3290426356/util/0.log" Nov 25 14:16:55 crc kubenswrapper[4702]: I1125 14:16:55.324836 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp_015eb990-26fa-467f-b645-fd3290426356/pull/0.log" Nov 25 14:16:55 crc kubenswrapper[4702]: I1125 14:16:55.357097 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992epcpsp_015eb990-26fa-467f-b645-fd3290426356/extract/0.log" Nov 25 14:16:55 crc kubenswrapper[4702]: I1125 14:16:55.504165 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p_bb1af68d-a7e9-428b-adcd-2a5b38d18fc7/util/0.log" Nov 25 14:16:55 crc kubenswrapper[4702]: I1125 14:16:55.730514 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p_bb1af68d-a7e9-428b-adcd-2a5b38d18fc7/pull/0.log" Nov 25 14:16:55 crc kubenswrapper[4702]: I1125 14:16:55.739376 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p_bb1af68d-a7e9-428b-adcd-2a5b38d18fc7/util/0.log" Nov 25 14:16:55 crc kubenswrapper[4702]: I1125 14:16:55.751846 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p_bb1af68d-a7e9-428b-adcd-2a5b38d18fc7/pull/0.log" Nov 25 14:16:55 crc kubenswrapper[4702]: I1125 14:16:55.887298 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p_bb1af68d-a7e9-428b-adcd-2a5b38d18fc7/util/0.log" Nov 25 14:16:55 crc kubenswrapper[4702]: I1125 14:16:55.915494 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p_bb1af68d-a7e9-428b-adcd-2a5b38d18fc7/extract/0.log" Nov 25 14:16:55 crc kubenswrapper[4702]: I1125 14:16:55.927829 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ac237df509a46455f4fbdee45046d5c81651870714971c8a8fa57fa00dqd82p_bb1af68d-a7e9-428b-adcd-2a5b38d18fc7/pull/0.log" Nov 25 14:16:55 crc kubenswrapper[4702]: I1125 14:16:55.959004 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9_8765d4d8-ce9f-4801-8d84-1badefac8c10/util/0.log" Nov 25 14:16:56 crc kubenswrapper[4702]: I1125 14:16:56.078634 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9_8765d4d8-ce9f-4801-8d84-1badefac8c10/util/0.log" Nov 25 14:16:56 crc kubenswrapper[4702]: I1125 14:16:56.106402 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9_8765d4d8-ce9f-4801-8d84-1badefac8c10/pull/0.log" Nov 25 14:16:56 crc kubenswrapper[4702]: I1125 14:16:56.106514 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9_8765d4d8-ce9f-4801-8d84-1badefac8c10/pull/0.log" Nov 25 14:16:56 crc kubenswrapper[4702]: I1125 14:16:56.306692 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9_8765d4d8-ce9f-4801-8d84-1badefac8c10/util/0.log" Nov 25 14:16:56 crc kubenswrapper[4702]: I1125 14:16:56.330985 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9_8765d4d8-ce9f-4801-8d84-1badefac8c10/extract/0.log" Nov 25 14:16:56 crc kubenswrapper[4702]: I1125 14:16:56.332167 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6v4jl9_8765d4d8-ce9f-4801-8d84-1badefac8c10/pull/0.log" Nov 25 14:16:56 crc kubenswrapper[4702]: I1125 14:16:56.414977 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7cdb5cbc57-h5pdv_7886ece7-80ed-4443-8e1e-522f0628421b/manager/0.log" Nov 25 14:16:56 crc kubenswrapper[4702]: I1125 14:16:56.509826 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-index-nxqqk_cc0a8640-1af2-47ac-a821-324511457b9d/registry-server/0.log" Nov 25 14:16:56 crc kubenswrapper[4702]: I1125 14:16:56.594030 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-664bc6f655-7r9vd_8611ff67-05cc-4795-83a1-a2be5b64dcd7/manager/0.log" Nov 25 14:16:56 crc kubenswrapper[4702]: I1125 14:16:56.706988 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-index-l2sm2_62409d84-e1be-4265-ae7b-b58face7988b/registry-server/0.log" Nov 25 14:16:56 crc kubenswrapper[4702]: I1125 14:16:56.773096 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-79898f8746-rfpp7_e13b9e67-df17-4851-a3c5-5037b994637d/kube-rbac-proxy/0.log" Nov 25 14:16:56 crc kubenswrapper[4702]: I1125 14:16:56.803677 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-79898f8746-rfpp7_e13b9e67-df17-4851-a3c5-5037b994637d/manager/0.log" Nov 25 14:16:56 crc kubenswrapper[4702]: I1125 14:16:56.954608 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-index-mgnpn_bf684cff-ed36-404a-8eba-acdb4cac164d/registry-server/0.log" Nov 25 14:16:56 crc kubenswrapper[4702]: I1125 14:16:56.970781 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-6b74b5c7-zfrjm_5b0aeb75-9843-4e68-9817-c1b0ac2062ea/manager/0.log" Nov 25 14:16:57 crc kubenswrapper[4702]: I1125 14:16:57.032674 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-index-rx4hj_ca7f19bd-2112-4f04-bd40-c8017f61804d/registry-server/0.log" Nov 25 14:16:57 crc kubenswrapper[4702]: I1125 14:16:57.145615 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-64cc8659b5-67ssg_9817b60d-cfac-44d3-b102-ec3a5670fbc2/manager/0.log" Nov 25 14:16:57 crc kubenswrapper[4702]: I1125 14:16:57.244448 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-index-4p8b5_0eaf7ea6-c46b-4af6-af99-006a4e6e06dd/registry-server/0.log" Nov 25 14:16:57 crc kubenswrapper[4702]: I1125 14:16:57.326661 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-779fc9694b-nzffd_3106c490-5a52-4dd1-9ef8-9ee2325a7d62/operator/0.log" Nov 25 14:16:57 crc kubenswrapper[4702]: I1125 14:16:57.370058 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-index-9wl6w_fc79fae5-b7b4-43b6-89ad-90008bf8dd57/registry-server/0.log" Nov 25 14:16:57 crc kubenswrapper[4702]: I1125 14:16:57.427750 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-7b9f5ffc8f-6zsw2_597dd90d-2ea6-4ea4-ad61-df59d4ca8944/manager/0.log" Nov 25 14:16:57 crc kubenswrapper[4702]: I1125 14:16:57.538686 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-index-9htnh_61fcd26c-69b0-4a87-951e-7a4f002acff9/registry-server/0.log" Nov 25 14:17:01 crc kubenswrapper[4702]: I1125 14:17:01.805623 4702 scope.go:117] "RemoveContainer" containerID="32273717263d7fcc99e8193a439dd976da66869c55bb4633d9314cacb6ab92a1" Nov 25 14:17:01 crc kubenswrapper[4702]: E1125 14:17:01.806138 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qqr76_openshift-machine-config-operator(3c491818-b0bb-4d82-a031-96e2dfac8c27)\"" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" Nov 25 14:17:11 crc kubenswrapper[4702]: I1125 14:17:11.428470 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-vjbjr_138d7b9f-9cae-436f-a06e-e3ccd0b9168d/control-plane-machine-set-operator/0.log" Nov 25 14:17:11 crc kubenswrapper[4702]: I1125 14:17:11.582382 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-9bq59_26e29bc4-82c6-47c1-8ffa-5e35e4b1c595/kube-rbac-proxy/0.log" Nov 25 14:17:11 crc kubenswrapper[4702]: I1125 14:17:11.606432 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-9bq59_26e29bc4-82c6-47c1-8ffa-5e35e4b1c595/machine-api-operator/0.log" Nov 25 14:17:14 crc kubenswrapper[4702]: I1125 14:17:14.801335 4702 scope.go:117] "RemoveContainer" containerID="32273717263d7fcc99e8193a439dd976da66869c55bb4633d9314cacb6ab92a1" Nov 25 14:17:14 crc kubenswrapper[4702]: E1125 14:17:14.801861 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qqr76_openshift-machine-config-operator(3c491818-b0bb-4d82-a031-96e2dfac8c27)\"" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" Nov 25 14:17:25 crc kubenswrapper[4702]: I1125 14:17:25.800938 4702 scope.go:117] "RemoveContainer" containerID="32273717263d7fcc99e8193a439dd976da66869c55bb4633d9314cacb6ab92a1" Nov 25 14:17:25 crc kubenswrapper[4702]: E1125 14:17:25.801863 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qqr76_openshift-machine-config-operator(3c491818-b0bb-4d82-a031-96e2dfac8c27)\"" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" Nov 25 14:17:27 crc kubenswrapper[4702]: I1125 14:17:27.565282 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-chvrp_d819dd6c-359a-4449-be38-1f0b18ae64d2/controller/0.log" Nov 25 14:17:27 crc kubenswrapper[4702]: I1125 14:17:27.578680 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-chvrp_d819dd6c-359a-4449-be38-1f0b18ae64d2/kube-rbac-proxy/0.log" Nov 25 14:17:27 crc kubenswrapper[4702]: I1125 14:17:27.723726 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hpc6n_5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3/cp-frr-files/0.log" Nov 25 14:17:27 crc kubenswrapper[4702]: I1125 14:17:27.869903 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hpc6n_5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3/cp-frr-files/0.log" Nov 25 14:17:27 crc kubenswrapper[4702]: I1125 14:17:27.870389 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hpc6n_5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3/cp-reloader/0.log" Nov 25 14:17:27 crc kubenswrapper[4702]: I1125 14:17:27.904561 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hpc6n_5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3/cp-metrics/0.log" Nov 25 14:17:27 crc kubenswrapper[4702]: I1125 14:17:27.918738 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hpc6n_5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3/cp-reloader/0.log" Nov 25 14:17:28 crc kubenswrapper[4702]: I1125 14:17:28.096627 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hpc6n_5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3/cp-reloader/0.log" Nov 25 14:17:28 crc kubenswrapper[4702]: I1125 14:17:28.117796 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hpc6n_5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3/cp-frr-files/0.log" Nov 25 14:17:28 crc kubenswrapper[4702]: I1125 14:17:28.141662 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hpc6n_5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3/cp-metrics/0.log" Nov 25 14:17:28 crc kubenswrapper[4702]: I1125 14:17:28.167663 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hpc6n_5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3/cp-metrics/0.log" Nov 25 14:17:28 crc kubenswrapper[4702]: I1125 14:17:28.315104 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hpc6n_5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3/cp-frr-files/0.log" Nov 25 14:17:28 crc kubenswrapper[4702]: I1125 14:17:28.315883 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hpc6n_5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3/cp-reloader/0.log" Nov 25 14:17:28 crc kubenswrapper[4702]: I1125 14:17:28.334920 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hpc6n_5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3/cp-metrics/0.log" Nov 25 14:17:28 crc kubenswrapper[4702]: I1125 14:17:28.355887 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hpc6n_5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3/controller/0.log" Nov 25 14:17:28 crc kubenswrapper[4702]: I1125 14:17:28.511270 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hpc6n_5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3/frr-metrics/0.log" Nov 25 14:17:28 crc kubenswrapper[4702]: I1125 14:17:28.539170 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hpc6n_5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3/kube-rbac-proxy/0.log" Nov 25 14:17:28 crc kubenswrapper[4702]: I1125 14:17:28.560571 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hpc6n_5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3/kube-rbac-proxy-frr/0.log" Nov 25 14:17:28 crc kubenswrapper[4702]: I1125 14:17:28.728560 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hpc6n_5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3/reloader/0.log" Nov 25 14:17:28 crc kubenswrapper[4702]: I1125 14:17:28.790471 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-btnfr_c0aec7c5-7588-4f42-a64a-9bbe4c07ea1c/frr-k8s-webhook-server/0.log" Nov 25 14:17:28 crc kubenswrapper[4702]: I1125 14:17:28.932799 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hpc6n_5e546573-b4d5-4c4f-ab0d-0f4ffa0837b3/frr/0.log" Nov 25 14:17:28 crc kubenswrapper[4702]: I1125 14:17:28.986608 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-7f98f4d986-wh2hc_422ee193-46e3-47d9-8452-29cc5a0fa7c8/manager/0.log" Nov 25 14:17:29 crc kubenswrapper[4702]: I1125 14:17:29.147849 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-757b88d98b-pkltf_28d9e60e-bc7b-4268-b855-93c5c98a19bb/webhook-server/0.log" Nov 25 14:17:29 crc kubenswrapper[4702]: I1125 14:17:29.209565 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-45qwr_4f16595d-125a-4e11-a745-abea073fe836/kube-rbac-proxy/0.log" Nov 25 14:17:29 crc kubenswrapper[4702]: I1125 14:17:29.372901 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-45qwr_4f16595d-125a-4e11-a745-abea073fe836/speaker/0.log" Nov 25 14:17:36 crc kubenswrapper[4702]: I1125 14:17:36.801067 4702 scope.go:117] "RemoveContainer" containerID="32273717263d7fcc99e8193a439dd976da66869c55bb4633d9314cacb6ab92a1" Nov 25 14:17:36 crc kubenswrapper[4702]: E1125 14:17:36.801791 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qqr76_openshift-machine-config-operator(3c491818-b0bb-4d82-a031-96e2dfac8c27)\"" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" Nov 25 14:17:42 crc kubenswrapper[4702]: I1125 14:17:42.991339 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_f17baa10-42c4-4620-9b3c-a5dde9f78229/mysql-bootstrap/0.log" Nov 25 14:17:43 crc kubenswrapper[4702]: I1125 14:17:43.098031 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_keystone-8c94d4449-cbvs8_791fc834-956f-4506-abf6-5d3695e23108/keystone-api/0.log" Nov 25 14:17:43 crc kubenswrapper[4702]: I1125 14:17:43.407207 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_f17baa10-42c4-4620-9b3c-a5dde9f78229/mysql-bootstrap/0.log" Nov 25 14:17:43 crc kubenswrapper[4702]: I1125 14:17:43.468025 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_f17baa10-42c4-4620-9b3c-a5dde9f78229/galera/0.log" Nov 25 14:17:43 crc kubenswrapper[4702]: I1125 14:17:43.618450 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_8a586ccc-2c61-418c-a84c-51730c841205/mysql-bootstrap/0.log" Nov 25 14:17:43 crc kubenswrapper[4702]: I1125 14:17:43.784063 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_8a586ccc-2c61-418c-a84c-51730c841205/galera/0.log" Nov 25 14:17:43 crc kubenswrapper[4702]: I1125 14:17:43.831525 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_8a586ccc-2c61-418c-a84c-51730c841205/mysql-bootstrap/0.log" Nov 25 14:17:44 crc kubenswrapper[4702]: I1125 14:17:44.020704 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_bbd4f8e2-cb5f-45ff-ad8f-0faef399b018/mysql-bootstrap/0.log" Nov 25 14:17:44 crc kubenswrapper[4702]: I1125 14:17:44.197069 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_bbd4f8e2-cb5f-45ff-ad8f-0faef399b018/mysql-bootstrap/0.log" Nov 25 14:17:44 crc kubenswrapper[4702]: I1125 14:17:44.260382 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_bbd4f8e2-cb5f-45ff-ad8f-0faef399b018/galera/0.log" Nov 25 14:17:44 crc kubenswrapper[4702]: I1125 14:17:44.333147 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_memcached-0_ac997ffc-32f2-4065-bc24-9982a8e306b6/memcached/0.log" Nov 25 14:17:44 crc kubenswrapper[4702]: I1125 14:17:44.409494 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstackclient_9366d936-2d43-4a0b-a8e6-4006afebd985/openstackclient/0.log" Nov 25 14:17:44 crc kubenswrapper[4702]: I1125 14:17:44.467994 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_3454fb60-87b9-45de-b8b2-13548d88cb1e/setup-container/0.log" Nov 25 14:17:44 crc kubenswrapper[4702]: I1125 14:17:44.669405 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_3454fb60-87b9-45de-b8b2-13548d88cb1e/setup-container/0.log" Nov 25 14:17:44 crc kubenswrapper[4702]: I1125 14:17:44.685818 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_3454fb60-87b9-45de-b8b2-13548d88cb1e/rabbitmq/0.log" Nov 25 14:17:44 crc kubenswrapper[4702]: I1125 14:17:44.729806 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-proxy-6bd58cfcf7-k4rxf_058da3b1-d848-4742-9454-f980618f76bd/proxy-httpd/0.log" Nov 25 14:17:44 crc kubenswrapper[4702]: I1125 14:17:44.888812 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-proxy-6bd58cfcf7-k4rxf_058da3b1-d848-4742-9454-f980618f76bd/proxy-server/0.log" Nov 25 14:17:44 crc kubenswrapper[4702]: I1125 14:17:44.920469 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-ring-rebalance-k5rd6_0f80b097-f79c-47fb-832d-950dd6045b4e/swift-ring-rebalance/0.log" Nov 25 14:17:45 crc kubenswrapper[4702]: I1125 14:17:45.074032 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_f332564e-cc0f-4d05-8816-139528cf0bb0/account-auditor/0.log" Nov 25 14:17:45 crc kubenswrapper[4702]: I1125 14:17:45.093353 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_f332564e-cc0f-4d05-8816-139528cf0bb0/account-reaper/0.log" Nov 25 14:17:45 crc kubenswrapper[4702]: I1125 14:17:45.117470 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_f332564e-cc0f-4d05-8816-139528cf0bb0/account-replicator/0.log" Nov 25 14:17:45 crc kubenswrapper[4702]: I1125 14:17:45.142502 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_f332564e-cc0f-4d05-8816-139528cf0bb0/account-server/0.log" Nov 25 14:17:45 crc kubenswrapper[4702]: I1125 14:17:45.267694 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_f332564e-cc0f-4d05-8816-139528cf0bb0/container-auditor/0.log" Nov 25 14:17:45 crc kubenswrapper[4702]: I1125 14:17:45.291327 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_f332564e-cc0f-4d05-8816-139528cf0bb0/container-replicator/0.log" Nov 25 14:17:45 crc kubenswrapper[4702]: I1125 14:17:45.333711 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_f332564e-cc0f-4d05-8816-139528cf0bb0/container-updater/0.log" Nov 25 14:17:45 crc kubenswrapper[4702]: I1125 14:17:45.351166 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_f332564e-cc0f-4d05-8816-139528cf0bb0/container-server/0.log" Nov 25 14:17:45 crc kubenswrapper[4702]: I1125 14:17:45.440283 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_f332564e-cc0f-4d05-8816-139528cf0bb0/object-auditor/0.log" Nov 25 14:17:45 crc kubenswrapper[4702]: I1125 14:17:45.487862 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_f332564e-cc0f-4d05-8816-139528cf0bb0/object-expirer/0.log" Nov 25 14:17:45 crc kubenswrapper[4702]: I1125 14:17:45.532222 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_f332564e-cc0f-4d05-8816-139528cf0bb0/object-server/0.log" Nov 25 14:17:45 crc kubenswrapper[4702]: I1125 14:17:45.543471 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_f332564e-cc0f-4d05-8816-139528cf0bb0/object-replicator/0.log" Nov 25 14:17:45 crc kubenswrapper[4702]: I1125 14:17:45.643649 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_f332564e-cc0f-4d05-8816-139528cf0bb0/object-updater/0.log" Nov 25 14:17:45 crc kubenswrapper[4702]: I1125 14:17:45.675415 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_f332564e-cc0f-4d05-8816-139528cf0bb0/rsync/0.log" Nov 25 14:17:45 crc kubenswrapper[4702]: I1125 14:17:45.701824 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_f332564e-cc0f-4d05-8816-139528cf0bb0/swift-recon-cron/0.log" Nov 25 14:17:47 crc kubenswrapper[4702]: I1125 14:17:47.802524 4702 scope.go:117] "RemoveContainer" containerID="32273717263d7fcc99e8193a439dd976da66869c55bb4633d9314cacb6ab92a1" Nov 25 14:17:47 crc kubenswrapper[4702]: E1125 14:17:47.804725 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qqr76_openshift-machine-config-operator(3c491818-b0bb-4d82-a031-96e2dfac8c27)\"" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" Nov 25 14:17:57 crc kubenswrapper[4702]: I1125 14:17:57.970617 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xlvf4_000d8b95-f22b-490b-b418-b8a070ed69f8/extract-utilities/0.log" Nov 25 14:17:58 crc kubenswrapper[4702]: I1125 14:17:58.112388 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xlvf4_000d8b95-f22b-490b-b418-b8a070ed69f8/extract-content/0.log" Nov 25 14:17:58 crc kubenswrapper[4702]: I1125 14:17:58.153799 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xlvf4_000d8b95-f22b-490b-b418-b8a070ed69f8/extract-utilities/0.log" Nov 25 14:17:58 crc kubenswrapper[4702]: I1125 14:17:58.154583 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xlvf4_000d8b95-f22b-490b-b418-b8a070ed69f8/extract-content/0.log" Nov 25 14:17:58 crc kubenswrapper[4702]: I1125 14:17:58.294788 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xlvf4_000d8b95-f22b-490b-b418-b8a070ed69f8/extract-content/0.log" Nov 25 14:17:58 crc kubenswrapper[4702]: I1125 14:17:58.297752 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xlvf4_000d8b95-f22b-490b-b418-b8a070ed69f8/extract-utilities/0.log" Nov 25 14:17:58 crc kubenswrapper[4702]: I1125 14:17:58.522191 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-r28xx_2b30ed31-5291-4679-89a2-ca54d9f477b8/extract-utilities/0.log" Nov 25 14:17:58 crc kubenswrapper[4702]: I1125 14:17:58.741328 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-r28xx_2b30ed31-5291-4679-89a2-ca54d9f477b8/extract-content/0.log" Nov 25 14:17:58 crc kubenswrapper[4702]: I1125 14:17:58.747921 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xlvf4_000d8b95-f22b-490b-b418-b8a070ed69f8/registry-server/0.log" Nov 25 14:17:58 crc kubenswrapper[4702]: I1125 14:17:58.772084 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-r28xx_2b30ed31-5291-4679-89a2-ca54d9f477b8/extract-utilities/0.log" Nov 25 14:17:58 crc kubenswrapper[4702]: I1125 14:17:58.779609 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-r28xx_2b30ed31-5291-4679-89a2-ca54d9f477b8/extract-content/0.log" Nov 25 14:17:58 crc kubenswrapper[4702]: I1125 14:17:58.909710 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-r28xx_2b30ed31-5291-4679-89a2-ca54d9f477b8/extract-content/0.log" Nov 25 14:17:58 crc kubenswrapper[4702]: I1125 14:17:58.934160 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-r28xx_2b30ed31-5291-4679-89a2-ca54d9f477b8/extract-utilities/0.log" Nov 25 14:17:59 crc kubenswrapper[4702]: I1125 14:17:59.163870 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7_97dcf135-5184-42db-b23e-b39dcbe81c99/util/0.log" Nov 25 14:17:59 crc kubenswrapper[4702]: I1125 14:17:59.424588 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7_97dcf135-5184-42db-b23e-b39dcbe81c99/pull/0.log" Nov 25 14:17:59 crc kubenswrapper[4702]: I1125 14:17:59.450124 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7_97dcf135-5184-42db-b23e-b39dcbe81c99/util/0.log" Nov 25 14:17:59 crc kubenswrapper[4702]: I1125 14:17:59.453667 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7_97dcf135-5184-42db-b23e-b39dcbe81c99/pull/0.log" Nov 25 14:17:59 crc kubenswrapper[4702]: I1125 14:17:59.536938 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-r28xx_2b30ed31-5291-4679-89a2-ca54d9f477b8/registry-server/0.log" Nov 25 14:17:59 crc kubenswrapper[4702]: I1125 14:17:59.612301 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7_97dcf135-5184-42db-b23e-b39dcbe81c99/util/0.log" Nov 25 14:17:59 crc kubenswrapper[4702]: I1125 14:17:59.651443 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7_97dcf135-5184-42db-b23e-b39dcbe81c99/extract/0.log" Nov 25 14:17:59 crc kubenswrapper[4702]: I1125 14:17:59.707756 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pj2r7_97dcf135-5184-42db-b23e-b39dcbe81c99/pull/0.log" Nov 25 14:17:59 crc kubenswrapper[4702]: I1125 14:17:59.794703 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-8mh98_b49d10e5-028c-4ad8-b03f-882744920b74/marketplace-operator/0.log" Nov 25 14:17:59 crc kubenswrapper[4702]: I1125 14:17:59.895719 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nhp97_ecf07980-291f-474e-9a27-9f050288a987/extract-utilities/0.log" Nov 25 14:18:00 crc kubenswrapper[4702]: I1125 14:18:00.069993 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nhp97_ecf07980-291f-474e-9a27-9f050288a987/extract-content/0.log" Nov 25 14:18:00 crc kubenswrapper[4702]: I1125 14:18:00.071079 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nhp97_ecf07980-291f-474e-9a27-9f050288a987/extract-utilities/0.log" Nov 25 14:18:00 crc kubenswrapper[4702]: I1125 14:18:00.075126 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nhp97_ecf07980-291f-474e-9a27-9f050288a987/extract-content/0.log" Nov 25 14:18:00 crc kubenswrapper[4702]: I1125 14:18:00.232376 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nhp97_ecf07980-291f-474e-9a27-9f050288a987/extract-utilities/0.log" Nov 25 14:18:00 crc kubenswrapper[4702]: I1125 14:18:00.276637 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nhp97_ecf07980-291f-474e-9a27-9f050288a987/extract-content/0.log" Nov 25 14:18:00 crc kubenswrapper[4702]: I1125 14:18:00.336795 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nhp97_ecf07980-291f-474e-9a27-9f050288a987/registry-server/0.log" Nov 25 14:18:00 crc kubenswrapper[4702]: I1125 14:18:00.430494 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5fmgv_afc764af-1603-432c-9993-d0fc66fbb4fa/extract-utilities/0.log" Nov 25 14:18:00 crc kubenswrapper[4702]: I1125 14:18:00.545887 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5fmgv_afc764af-1603-432c-9993-d0fc66fbb4fa/extract-utilities/0.log" Nov 25 14:18:00 crc kubenswrapper[4702]: I1125 14:18:00.568834 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5fmgv_afc764af-1603-432c-9993-d0fc66fbb4fa/extract-content/0.log" Nov 25 14:18:00 crc kubenswrapper[4702]: I1125 14:18:00.644810 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5fmgv_afc764af-1603-432c-9993-d0fc66fbb4fa/extract-content/0.log" Nov 25 14:18:00 crc kubenswrapper[4702]: I1125 14:18:00.771575 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5fmgv_afc764af-1603-432c-9993-d0fc66fbb4fa/extract-utilities/0.log" Nov 25 14:18:00 crc kubenswrapper[4702]: I1125 14:18:00.823269 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5fmgv_afc764af-1603-432c-9993-d0fc66fbb4fa/extract-content/0.log" Nov 25 14:18:01 crc kubenswrapper[4702]: I1125 14:18:01.221408 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5fmgv_afc764af-1603-432c-9993-d0fc66fbb4fa/registry-server/0.log" Nov 25 14:18:01 crc kubenswrapper[4702]: I1125 14:18:01.801099 4702 scope.go:117] "RemoveContainer" containerID="32273717263d7fcc99e8193a439dd976da66869c55bb4633d9314cacb6ab92a1" Nov 25 14:18:01 crc kubenswrapper[4702]: E1125 14:18:01.801445 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qqr76_openshift-machine-config-operator(3c491818-b0bb-4d82-a031-96e2dfac8c27)\"" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" Nov 25 14:18:12 crc kubenswrapper[4702]: I1125 14:18:12.800610 4702 scope.go:117] "RemoveContainer" containerID="32273717263d7fcc99e8193a439dd976da66869c55bb4633d9314cacb6ab92a1" Nov 25 14:18:12 crc kubenswrapper[4702]: E1125 14:18:12.801373 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qqr76_openshift-machine-config-operator(3c491818-b0bb-4d82-a031-96e2dfac8c27)\"" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" Nov 25 14:18:23 crc kubenswrapper[4702]: I1125 14:18:23.801130 4702 scope.go:117] "RemoveContainer" containerID="32273717263d7fcc99e8193a439dd976da66869c55bb4633d9314cacb6ab92a1" Nov 25 14:18:23 crc kubenswrapper[4702]: E1125 14:18:23.801748 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qqr76_openshift-machine-config-operator(3c491818-b0bb-4d82-a031-96e2dfac8c27)\"" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" Nov 25 14:18:36 crc kubenswrapper[4702]: I1125 14:18:36.801611 4702 scope.go:117] "RemoveContainer" containerID="32273717263d7fcc99e8193a439dd976da66869c55bb4633d9314cacb6ab92a1" Nov 25 14:18:36 crc kubenswrapper[4702]: E1125 14:18:36.802907 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qqr76_openshift-machine-config-operator(3c491818-b0bb-4d82-a031-96e2dfac8c27)\"" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" Nov 25 14:18:48 crc kubenswrapper[4702]: I1125 14:18:48.801921 4702 scope.go:117] "RemoveContainer" containerID="32273717263d7fcc99e8193a439dd976da66869c55bb4633d9314cacb6ab92a1" Nov 25 14:18:48 crc kubenswrapper[4702]: E1125 14:18:48.802488 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qqr76_openshift-machine-config-operator(3c491818-b0bb-4d82-a031-96e2dfac8c27)\"" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" Nov 25 14:19:01 crc kubenswrapper[4702]: I1125 14:19:01.805507 4702 scope.go:117] "RemoveContainer" containerID="32273717263d7fcc99e8193a439dd976da66869c55bb4633d9314cacb6ab92a1" Nov 25 14:19:01 crc kubenswrapper[4702]: E1125 14:19:01.806116 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qqr76_openshift-machine-config-operator(3c491818-b0bb-4d82-a031-96e2dfac8c27)\"" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" Nov 25 14:19:08 crc kubenswrapper[4702]: I1125 14:19:08.628554 4702 generic.go:334] "Generic (PLEG): container finished" podID="5bd1aa0b-893d-4550-85eb-3ee257c7b355" containerID="d89d654239cd9f7f93a9910cdfea090141ef0f2cb7223575974745f290431aa1" exitCode=0 Nov 25 14:19:08 crc kubenswrapper[4702]: I1125 14:19:08.628665 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-k9pzw/must-gather-wsn6h" event={"ID":"5bd1aa0b-893d-4550-85eb-3ee257c7b355","Type":"ContainerDied","Data":"d89d654239cd9f7f93a9910cdfea090141ef0f2cb7223575974745f290431aa1"} Nov 25 14:19:08 crc kubenswrapper[4702]: I1125 14:19:08.629866 4702 scope.go:117] "RemoveContainer" containerID="d89d654239cd9f7f93a9910cdfea090141ef0f2cb7223575974745f290431aa1" Nov 25 14:19:08 crc kubenswrapper[4702]: I1125 14:19:08.857053 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-k9pzw_must-gather-wsn6h_5bd1aa0b-893d-4550-85eb-3ee257c7b355/gather/0.log" Nov 25 14:19:16 crc kubenswrapper[4702]: I1125 14:19:16.274774 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-k9pzw/must-gather-wsn6h"] Nov 25 14:19:16 crc kubenswrapper[4702]: I1125 14:19:16.275823 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-k9pzw/must-gather-wsn6h" podUID="5bd1aa0b-893d-4550-85eb-3ee257c7b355" containerName="copy" containerID="cri-o://343a84620bdc6ee504942e6a904d4bf9437e49094c13f6176cabb793a0524bbb" gracePeriod=2 Nov 25 14:19:16 crc kubenswrapper[4702]: I1125 14:19:16.282738 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-k9pzw/must-gather-wsn6h"] Nov 25 14:19:16 crc kubenswrapper[4702]: I1125 14:19:16.690095 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-k9pzw_must-gather-wsn6h_5bd1aa0b-893d-4550-85eb-3ee257c7b355/copy/0.log" Nov 25 14:19:16 crc kubenswrapper[4702]: I1125 14:19:16.690752 4702 generic.go:334] "Generic (PLEG): container finished" podID="5bd1aa0b-893d-4550-85eb-3ee257c7b355" containerID="343a84620bdc6ee504942e6a904d4bf9437e49094c13f6176cabb793a0524bbb" exitCode=143 Nov 25 14:19:16 crc kubenswrapper[4702]: I1125 14:19:16.690789 4702 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca3a19068192db6cbdc519cda282633d26fef2399bc1849ffa91f943101653f6" Nov 25 14:19:16 crc kubenswrapper[4702]: I1125 14:19:16.720872 4702 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-k9pzw_must-gather-wsn6h_5bd1aa0b-893d-4550-85eb-3ee257c7b355/copy/0.log" Nov 25 14:19:16 crc kubenswrapper[4702]: I1125 14:19:16.721317 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k9pzw/must-gather-wsn6h" Nov 25 14:19:16 crc kubenswrapper[4702]: I1125 14:19:16.802216 4702 scope.go:117] "RemoveContainer" containerID="32273717263d7fcc99e8193a439dd976da66869c55bb4633d9314cacb6ab92a1" Nov 25 14:19:16 crc kubenswrapper[4702]: E1125 14:19:16.802461 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qqr76_openshift-machine-config-operator(3c491818-b0bb-4d82-a031-96e2dfac8c27)\"" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" Nov 25 14:19:16 crc kubenswrapper[4702]: I1125 14:19:16.915776 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hksw6\" (UniqueName: \"kubernetes.io/projected/5bd1aa0b-893d-4550-85eb-3ee257c7b355-kube-api-access-hksw6\") pod \"5bd1aa0b-893d-4550-85eb-3ee257c7b355\" (UID: \"5bd1aa0b-893d-4550-85eb-3ee257c7b355\") " Nov 25 14:19:16 crc kubenswrapper[4702]: I1125 14:19:16.916601 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5bd1aa0b-893d-4550-85eb-3ee257c7b355-must-gather-output\") pod \"5bd1aa0b-893d-4550-85eb-3ee257c7b355\" (UID: \"5bd1aa0b-893d-4550-85eb-3ee257c7b355\") " Nov 25 14:19:16 crc kubenswrapper[4702]: I1125 14:19:16.922218 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bd1aa0b-893d-4550-85eb-3ee257c7b355-kube-api-access-hksw6" (OuterVolumeSpecName: "kube-api-access-hksw6") pod "5bd1aa0b-893d-4550-85eb-3ee257c7b355" (UID: "5bd1aa0b-893d-4550-85eb-3ee257c7b355"). InnerVolumeSpecName "kube-api-access-hksw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:19:16 crc kubenswrapper[4702]: I1125 14:19:16.984873 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bd1aa0b-893d-4550-85eb-3ee257c7b355-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "5bd1aa0b-893d-4550-85eb-3ee257c7b355" (UID: "5bd1aa0b-893d-4550-85eb-3ee257c7b355"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:19:17 crc kubenswrapper[4702]: I1125 14:19:17.018170 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hksw6\" (UniqueName: \"kubernetes.io/projected/5bd1aa0b-893d-4550-85eb-3ee257c7b355-kube-api-access-hksw6\") on node \"crc\" DevicePath \"\"" Nov 25 14:19:17 crc kubenswrapper[4702]: I1125 14:19:17.018203 4702 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5bd1aa0b-893d-4550-85eb-3ee257c7b355-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 25 14:19:17 crc kubenswrapper[4702]: I1125 14:19:17.696821 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k9pzw/must-gather-wsn6h" Nov 25 14:19:17 crc kubenswrapper[4702]: I1125 14:19:17.809128 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bd1aa0b-893d-4550-85eb-3ee257c7b355" path="/var/lib/kubelet/pods/5bd1aa0b-893d-4550-85eb-3ee257c7b355/volumes" Nov 25 14:19:30 crc kubenswrapper[4702]: I1125 14:19:30.800711 4702 scope.go:117] "RemoveContainer" containerID="32273717263d7fcc99e8193a439dd976da66869c55bb4633d9314cacb6ab92a1" Nov 25 14:19:30 crc kubenswrapper[4702]: E1125 14:19:30.801456 4702 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qqr76_openshift-machine-config-operator(3c491818-b0bb-4d82-a031-96e2dfac8c27)\"" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" Nov 25 14:19:45 crc kubenswrapper[4702]: I1125 14:19:45.818651 4702 scope.go:117] "RemoveContainer" containerID="32273717263d7fcc99e8193a439dd976da66869c55bb4633d9314cacb6ab92a1" Nov 25 14:19:46 crc kubenswrapper[4702]: I1125 14:19:46.917766 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" event={"ID":"3c491818-b0bb-4d82-a031-96e2dfac8c27","Type":"ContainerStarted","Data":"9f8fe70e21680c908554f1b846512eb5113ef4d904ef8a32b1e094b610045301"} Nov 25 14:21:38 crc kubenswrapper[4702]: I1125 14:21:38.278984 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9dlnl"] Nov 25 14:21:38 crc kubenswrapper[4702]: E1125 14:21:38.279872 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bd1aa0b-893d-4550-85eb-3ee257c7b355" containerName="copy" Nov 25 14:21:38 crc kubenswrapper[4702]: I1125 14:21:38.279884 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bd1aa0b-893d-4550-85eb-3ee257c7b355" containerName="copy" Nov 25 14:21:38 crc kubenswrapper[4702]: E1125 14:21:38.279902 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bd1aa0b-893d-4550-85eb-3ee257c7b355" containerName="gather" Nov 25 14:21:38 crc kubenswrapper[4702]: I1125 14:21:38.279908 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bd1aa0b-893d-4550-85eb-3ee257c7b355" containerName="gather" Nov 25 14:21:38 crc kubenswrapper[4702]: I1125 14:21:38.280057 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bd1aa0b-893d-4550-85eb-3ee257c7b355" containerName="copy" Nov 25 14:21:38 crc kubenswrapper[4702]: I1125 14:21:38.280079 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bd1aa0b-893d-4550-85eb-3ee257c7b355" containerName="gather" Nov 25 14:21:38 crc kubenswrapper[4702]: I1125 14:21:38.281276 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9dlnl" Nov 25 14:21:38 crc kubenswrapper[4702]: I1125 14:21:38.296275 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9dlnl"] Nov 25 14:21:38 crc kubenswrapper[4702]: I1125 14:21:38.407849 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35ede4f2-f326-4051-8404-71125cf9576f-catalog-content\") pod \"certified-operators-9dlnl\" (UID: \"35ede4f2-f326-4051-8404-71125cf9576f\") " pod="openshift-marketplace/certified-operators-9dlnl" Nov 25 14:21:38 crc kubenswrapper[4702]: I1125 14:21:38.407988 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35ede4f2-f326-4051-8404-71125cf9576f-utilities\") pod \"certified-operators-9dlnl\" (UID: \"35ede4f2-f326-4051-8404-71125cf9576f\") " pod="openshift-marketplace/certified-operators-9dlnl" Nov 25 14:21:38 crc kubenswrapper[4702]: I1125 14:21:38.408013 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkf5p\" (UniqueName: \"kubernetes.io/projected/35ede4f2-f326-4051-8404-71125cf9576f-kube-api-access-xkf5p\") pod \"certified-operators-9dlnl\" (UID: \"35ede4f2-f326-4051-8404-71125cf9576f\") " pod="openshift-marketplace/certified-operators-9dlnl" Nov 25 14:21:38 crc kubenswrapper[4702]: I1125 14:21:38.527832 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35ede4f2-f326-4051-8404-71125cf9576f-catalog-content\") pod \"certified-operators-9dlnl\" (UID: \"35ede4f2-f326-4051-8404-71125cf9576f\") " pod="openshift-marketplace/certified-operators-9dlnl" Nov 25 14:21:38 crc kubenswrapper[4702]: I1125 14:21:38.528045 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35ede4f2-f326-4051-8404-71125cf9576f-utilities\") pod \"certified-operators-9dlnl\" (UID: \"35ede4f2-f326-4051-8404-71125cf9576f\") " pod="openshift-marketplace/certified-operators-9dlnl" Nov 25 14:21:38 crc kubenswrapper[4702]: I1125 14:21:38.528085 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkf5p\" (UniqueName: \"kubernetes.io/projected/35ede4f2-f326-4051-8404-71125cf9576f-kube-api-access-xkf5p\") pod \"certified-operators-9dlnl\" (UID: \"35ede4f2-f326-4051-8404-71125cf9576f\") " pod="openshift-marketplace/certified-operators-9dlnl" Nov 25 14:21:38 crc kubenswrapper[4702]: I1125 14:21:38.528343 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35ede4f2-f326-4051-8404-71125cf9576f-catalog-content\") pod \"certified-operators-9dlnl\" (UID: \"35ede4f2-f326-4051-8404-71125cf9576f\") " pod="openshift-marketplace/certified-operators-9dlnl" Nov 25 14:21:38 crc kubenswrapper[4702]: I1125 14:21:38.528392 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35ede4f2-f326-4051-8404-71125cf9576f-utilities\") pod \"certified-operators-9dlnl\" (UID: \"35ede4f2-f326-4051-8404-71125cf9576f\") " pod="openshift-marketplace/certified-operators-9dlnl" Nov 25 14:21:38 crc kubenswrapper[4702]: I1125 14:21:38.555205 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkf5p\" (UniqueName: \"kubernetes.io/projected/35ede4f2-f326-4051-8404-71125cf9576f-kube-api-access-xkf5p\") pod \"certified-operators-9dlnl\" (UID: \"35ede4f2-f326-4051-8404-71125cf9576f\") " pod="openshift-marketplace/certified-operators-9dlnl" Nov 25 14:21:38 crc kubenswrapper[4702]: I1125 14:21:38.604421 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9dlnl" Nov 25 14:21:38 crc kubenswrapper[4702]: I1125 14:21:38.870871 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9dlnl"] Nov 25 14:21:39 crc kubenswrapper[4702]: I1125 14:21:39.475860 4702 generic.go:334] "Generic (PLEG): container finished" podID="35ede4f2-f326-4051-8404-71125cf9576f" containerID="9daf9eb36d2112372fb575044b0096ea73066125b74e678a02627d3facae092e" exitCode=0 Nov 25 14:21:39 crc kubenswrapper[4702]: I1125 14:21:39.475918 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9dlnl" event={"ID":"35ede4f2-f326-4051-8404-71125cf9576f","Type":"ContainerDied","Data":"9daf9eb36d2112372fb575044b0096ea73066125b74e678a02627d3facae092e"} Nov 25 14:21:39 crc kubenswrapper[4702]: I1125 14:21:39.476126 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9dlnl" event={"ID":"35ede4f2-f326-4051-8404-71125cf9576f","Type":"ContainerStarted","Data":"8f7505c63e607d3b41276212d4a355093390c821b6aa005791311f71bea8fdaa"} Nov 25 14:21:39 crc kubenswrapper[4702]: I1125 14:21:39.477950 4702 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 14:21:41 crc kubenswrapper[4702]: I1125 14:21:41.501594 4702 generic.go:334] "Generic (PLEG): container finished" podID="35ede4f2-f326-4051-8404-71125cf9576f" containerID="5e1129b93328c86f2b344a2352dbdd456e36c82e554ee73dc22a99f3fe77d9e0" exitCode=0 Nov 25 14:21:41 crc kubenswrapper[4702]: I1125 14:21:41.501667 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9dlnl" event={"ID":"35ede4f2-f326-4051-8404-71125cf9576f","Type":"ContainerDied","Data":"5e1129b93328c86f2b344a2352dbdd456e36c82e554ee73dc22a99f3fe77d9e0"} Nov 25 14:21:42 crc kubenswrapper[4702]: I1125 14:21:42.515757 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9dlnl" event={"ID":"35ede4f2-f326-4051-8404-71125cf9576f","Type":"ContainerStarted","Data":"5b0c290546ef5970de20637a18ae9ce8a2ddbeb69d40c42c6fce011ce54ceb3a"} Nov 25 14:21:42 crc kubenswrapper[4702]: I1125 14:21:42.536051 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9dlnl" podStartSLOduration=2.045558495 podStartE2EDuration="4.536028918s" podCreationTimestamp="2025-11-25 14:21:38 +0000 UTC" firstStartedPulling="2025-11-25 14:21:39.477609596 +0000 UTC m=+2108.640260046" lastFinishedPulling="2025-11-25 14:21:41.968080009 +0000 UTC m=+2111.130730469" observedRunningTime="2025-11-25 14:21:42.5352782 +0000 UTC m=+2111.697928650" watchObservedRunningTime="2025-11-25 14:21:42.536028918 +0000 UTC m=+2111.698679368" Nov 25 14:21:48 crc kubenswrapper[4702]: I1125 14:21:48.605367 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9dlnl" Nov 25 14:21:48 crc kubenswrapper[4702]: I1125 14:21:48.606507 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9dlnl" Nov 25 14:21:48 crc kubenswrapper[4702]: I1125 14:21:48.655943 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9dlnl" Nov 25 14:21:49 crc kubenswrapper[4702]: I1125 14:21:49.643445 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9dlnl" Nov 25 14:21:49 crc kubenswrapper[4702]: I1125 14:21:49.712886 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9dlnl"] Nov 25 14:21:51 crc kubenswrapper[4702]: I1125 14:21:51.598463 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9dlnl" podUID="35ede4f2-f326-4051-8404-71125cf9576f" containerName="registry-server" containerID="cri-o://5b0c290546ef5970de20637a18ae9ce8a2ddbeb69d40c42c6fce011ce54ceb3a" gracePeriod=2 Nov 25 14:21:51 crc kubenswrapper[4702]: I1125 14:21:51.976275 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9dlnl" Nov 25 14:21:52 crc kubenswrapper[4702]: I1125 14:21:52.066732 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkf5p\" (UniqueName: \"kubernetes.io/projected/35ede4f2-f326-4051-8404-71125cf9576f-kube-api-access-xkf5p\") pod \"35ede4f2-f326-4051-8404-71125cf9576f\" (UID: \"35ede4f2-f326-4051-8404-71125cf9576f\") " Nov 25 14:21:52 crc kubenswrapper[4702]: I1125 14:21:52.072921 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35ede4f2-f326-4051-8404-71125cf9576f-kube-api-access-xkf5p" (OuterVolumeSpecName: "kube-api-access-xkf5p") pod "35ede4f2-f326-4051-8404-71125cf9576f" (UID: "35ede4f2-f326-4051-8404-71125cf9576f"). InnerVolumeSpecName "kube-api-access-xkf5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:21:52 crc kubenswrapper[4702]: I1125 14:21:52.168098 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35ede4f2-f326-4051-8404-71125cf9576f-catalog-content\") pod \"35ede4f2-f326-4051-8404-71125cf9576f\" (UID: \"35ede4f2-f326-4051-8404-71125cf9576f\") " Nov 25 14:21:52 crc kubenswrapper[4702]: I1125 14:21:52.168292 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35ede4f2-f326-4051-8404-71125cf9576f-utilities\") pod \"35ede4f2-f326-4051-8404-71125cf9576f\" (UID: \"35ede4f2-f326-4051-8404-71125cf9576f\") " Nov 25 14:21:52 crc kubenswrapper[4702]: I1125 14:21:52.169043 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkf5p\" (UniqueName: \"kubernetes.io/projected/35ede4f2-f326-4051-8404-71125cf9576f-kube-api-access-xkf5p\") on node \"crc\" DevicePath \"\"" Nov 25 14:21:52 crc kubenswrapper[4702]: I1125 14:21:52.169308 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35ede4f2-f326-4051-8404-71125cf9576f-utilities" (OuterVolumeSpecName: "utilities") pod "35ede4f2-f326-4051-8404-71125cf9576f" (UID: "35ede4f2-f326-4051-8404-71125cf9576f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:21:52 crc kubenswrapper[4702]: I1125 14:21:52.223104 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35ede4f2-f326-4051-8404-71125cf9576f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "35ede4f2-f326-4051-8404-71125cf9576f" (UID: "35ede4f2-f326-4051-8404-71125cf9576f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:21:52 crc kubenswrapper[4702]: I1125 14:21:52.269727 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35ede4f2-f326-4051-8404-71125cf9576f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 14:21:52 crc kubenswrapper[4702]: I1125 14:21:52.269987 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35ede4f2-f326-4051-8404-71125cf9576f-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 14:21:52 crc kubenswrapper[4702]: I1125 14:21:52.609291 4702 generic.go:334] "Generic (PLEG): container finished" podID="35ede4f2-f326-4051-8404-71125cf9576f" containerID="5b0c290546ef5970de20637a18ae9ce8a2ddbeb69d40c42c6fce011ce54ceb3a" exitCode=0 Nov 25 14:21:52 crc kubenswrapper[4702]: I1125 14:21:52.609348 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9dlnl" event={"ID":"35ede4f2-f326-4051-8404-71125cf9576f","Type":"ContainerDied","Data":"5b0c290546ef5970de20637a18ae9ce8a2ddbeb69d40c42c6fce011ce54ceb3a"} Nov 25 14:21:52 crc kubenswrapper[4702]: I1125 14:21:52.609366 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9dlnl" Nov 25 14:21:52 crc kubenswrapper[4702]: I1125 14:21:52.609382 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9dlnl" event={"ID":"35ede4f2-f326-4051-8404-71125cf9576f","Type":"ContainerDied","Data":"8f7505c63e607d3b41276212d4a355093390c821b6aa005791311f71bea8fdaa"} Nov 25 14:21:52 crc kubenswrapper[4702]: I1125 14:21:52.609406 4702 scope.go:117] "RemoveContainer" containerID="5b0c290546ef5970de20637a18ae9ce8a2ddbeb69d40c42c6fce011ce54ceb3a" Nov 25 14:21:52 crc kubenswrapper[4702]: I1125 14:21:52.639211 4702 scope.go:117] "RemoveContainer" containerID="5e1129b93328c86f2b344a2352dbdd456e36c82e554ee73dc22a99f3fe77d9e0" Nov 25 14:21:52 crc kubenswrapper[4702]: I1125 14:21:52.650871 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9dlnl"] Nov 25 14:21:52 crc kubenswrapper[4702]: I1125 14:21:52.661846 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9dlnl"] Nov 25 14:21:52 crc kubenswrapper[4702]: I1125 14:21:52.695852 4702 scope.go:117] "RemoveContainer" containerID="9daf9eb36d2112372fb575044b0096ea73066125b74e678a02627d3facae092e" Nov 25 14:21:52 crc kubenswrapper[4702]: I1125 14:21:52.716451 4702 scope.go:117] "RemoveContainer" containerID="5b0c290546ef5970de20637a18ae9ce8a2ddbeb69d40c42c6fce011ce54ceb3a" Nov 25 14:21:52 crc kubenswrapper[4702]: E1125 14:21:52.716993 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b0c290546ef5970de20637a18ae9ce8a2ddbeb69d40c42c6fce011ce54ceb3a\": container with ID starting with 5b0c290546ef5970de20637a18ae9ce8a2ddbeb69d40c42c6fce011ce54ceb3a not found: ID does not exist" containerID="5b0c290546ef5970de20637a18ae9ce8a2ddbeb69d40c42c6fce011ce54ceb3a" Nov 25 14:21:52 crc kubenswrapper[4702]: I1125 14:21:52.717050 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b0c290546ef5970de20637a18ae9ce8a2ddbeb69d40c42c6fce011ce54ceb3a"} err="failed to get container status \"5b0c290546ef5970de20637a18ae9ce8a2ddbeb69d40c42c6fce011ce54ceb3a\": rpc error: code = NotFound desc = could not find container \"5b0c290546ef5970de20637a18ae9ce8a2ddbeb69d40c42c6fce011ce54ceb3a\": container with ID starting with 5b0c290546ef5970de20637a18ae9ce8a2ddbeb69d40c42c6fce011ce54ceb3a not found: ID does not exist" Nov 25 14:21:52 crc kubenswrapper[4702]: I1125 14:21:52.717085 4702 scope.go:117] "RemoveContainer" containerID="5e1129b93328c86f2b344a2352dbdd456e36c82e554ee73dc22a99f3fe77d9e0" Nov 25 14:21:52 crc kubenswrapper[4702]: E1125 14:21:52.717521 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e1129b93328c86f2b344a2352dbdd456e36c82e554ee73dc22a99f3fe77d9e0\": container with ID starting with 5e1129b93328c86f2b344a2352dbdd456e36c82e554ee73dc22a99f3fe77d9e0 not found: ID does not exist" containerID="5e1129b93328c86f2b344a2352dbdd456e36c82e554ee73dc22a99f3fe77d9e0" Nov 25 14:21:52 crc kubenswrapper[4702]: I1125 14:21:52.717588 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e1129b93328c86f2b344a2352dbdd456e36c82e554ee73dc22a99f3fe77d9e0"} err="failed to get container status \"5e1129b93328c86f2b344a2352dbdd456e36c82e554ee73dc22a99f3fe77d9e0\": rpc error: code = NotFound desc = could not find container \"5e1129b93328c86f2b344a2352dbdd456e36c82e554ee73dc22a99f3fe77d9e0\": container with ID starting with 5e1129b93328c86f2b344a2352dbdd456e36c82e554ee73dc22a99f3fe77d9e0 not found: ID does not exist" Nov 25 14:21:52 crc kubenswrapper[4702]: I1125 14:21:52.717617 4702 scope.go:117] "RemoveContainer" containerID="9daf9eb36d2112372fb575044b0096ea73066125b74e678a02627d3facae092e" Nov 25 14:21:52 crc kubenswrapper[4702]: E1125 14:21:52.717868 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9daf9eb36d2112372fb575044b0096ea73066125b74e678a02627d3facae092e\": container with ID starting with 9daf9eb36d2112372fb575044b0096ea73066125b74e678a02627d3facae092e not found: ID does not exist" containerID="9daf9eb36d2112372fb575044b0096ea73066125b74e678a02627d3facae092e" Nov 25 14:21:52 crc kubenswrapper[4702]: I1125 14:21:52.717904 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9daf9eb36d2112372fb575044b0096ea73066125b74e678a02627d3facae092e"} err="failed to get container status \"9daf9eb36d2112372fb575044b0096ea73066125b74e678a02627d3facae092e\": rpc error: code = NotFound desc = could not find container \"9daf9eb36d2112372fb575044b0096ea73066125b74e678a02627d3facae092e\": container with ID starting with 9daf9eb36d2112372fb575044b0096ea73066125b74e678a02627d3facae092e not found: ID does not exist" Nov 25 14:21:53 crc kubenswrapper[4702]: I1125 14:21:53.809758 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35ede4f2-f326-4051-8404-71125cf9576f" path="/var/lib/kubelet/pods/35ede4f2-f326-4051-8404-71125cf9576f/volumes" Nov 25 14:21:58 crc kubenswrapper[4702]: I1125 14:21:58.286986 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-k4rxf" podUID="058da3b1-d848-4742-9454-f980618f76bd" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 502" Nov 25 14:22:00 crc kubenswrapper[4702]: I1125 14:22:00.045827 4702 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-q8m72"] Nov 25 14:22:00 crc kubenswrapper[4702]: E1125 14:22:00.046717 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35ede4f2-f326-4051-8404-71125cf9576f" containerName="extract-utilities" Nov 25 14:22:00 crc kubenswrapper[4702]: I1125 14:22:00.046738 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="35ede4f2-f326-4051-8404-71125cf9576f" containerName="extract-utilities" Nov 25 14:22:00 crc kubenswrapper[4702]: E1125 14:22:00.046769 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35ede4f2-f326-4051-8404-71125cf9576f" containerName="extract-content" Nov 25 14:22:00 crc kubenswrapper[4702]: I1125 14:22:00.046783 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="35ede4f2-f326-4051-8404-71125cf9576f" containerName="extract-content" Nov 25 14:22:00 crc kubenswrapper[4702]: E1125 14:22:00.046815 4702 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35ede4f2-f326-4051-8404-71125cf9576f" containerName="registry-server" Nov 25 14:22:00 crc kubenswrapper[4702]: I1125 14:22:00.046830 4702 state_mem.go:107] "Deleted CPUSet assignment" podUID="35ede4f2-f326-4051-8404-71125cf9576f" containerName="registry-server" Nov 25 14:22:00 crc kubenswrapper[4702]: I1125 14:22:00.047071 4702 memory_manager.go:354] "RemoveStaleState removing state" podUID="35ede4f2-f326-4051-8404-71125cf9576f" containerName="registry-server" Nov 25 14:22:00 crc kubenswrapper[4702]: I1125 14:22:00.048745 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q8m72" Nov 25 14:22:00 crc kubenswrapper[4702]: I1125 14:22:00.060082 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q8m72"] Nov 25 14:22:00 crc kubenswrapper[4702]: I1125 14:22:00.110796 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/314da235-33f1-44f7-b9f0-3bc1af92879f-utilities\") pod \"redhat-marketplace-q8m72\" (UID: \"314da235-33f1-44f7-b9f0-3bc1af92879f\") " pod="openshift-marketplace/redhat-marketplace-q8m72" Nov 25 14:22:00 crc kubenswrapper[4702]: I1125 14:22:00.110971 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/314da235-33f1-44f7-b9f0-3bc1af92879f-catalog-content\") pod \"redhat-marketplace-q8m72\" (UID: \"314da235-33f1-44f7-b9f0-3bc1af92879f\") " pod="openshift-marketplace/redhat-marketplace-q8m72" Nov 25 14:22:00 crc kubenswrapper[4702]: I1125 14:22:00.111017 4702 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnwhv\" (UniqueName: \"kubernetes.io/projected/314da235-33f1-44f7-b9f0-3bc1af92879f-kube-api-access-qnwhv\") pod \"redhat-marketplace-q8m72\" (UID: \"314da235-33f1-44f7-b9f0-3bc1af92879f\") " pod="openshift-marketplace/redhat-marketplace-q8m72" Nov 25 14:22:00 crc kubenswrapper[4702]: I1125 14:22:00.212464 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/314da235-33f1-44f7-b9f0-3bc1af92879f-catalog-content\") pod \"redhat-marketplace-q8m72\" (UID: \"314da235-33f1-44f7-b9f0-3bc1af92879f\") " pod="openshift-marketplace/redhat-marketplace-q8m72" Nov 25 14:22:00 crc kubenswrapper[4702]: I1125 14:22:00.212521 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnwhv\" (UniqueName: \"kubernetes.io/projected/314da235-33f1-44f7-b9f0-3bc1af92879f-kube-api-access-qnwhv\") pod \"redhat-marketplace-q8m72\" (UID: \"314da235-33f1-44f7-b9f0-3bc1af92879f\") " pod="openshift-marketplace/redhat-marketplace-q8m72" Nov 25 14:22:00 crc kubenswrapper[4702]: I1125 14:22:00.212546 4702 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/314da235-33f1-44f7-b9f0-3bc1af92879f-utilities\") pod \"redhat-marketplace-q8m72\" (UID: \"314da235-33f1-44f7-b9f0-3bc1af92879f\") " pod="openshift-marketplace/redhat-marketplace-q8m72" Nov 25 14:22:00 crc kubenswrapper[4702]: I1125 14:22:00.212919 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/314da235-33f1-44f7-b9f0-3bc1af92879f-catalog-content\") pod \"redhat-marketplace-q8m72\" (UID: \"314da235-33f1-44f7-b9f0-3bc1af92879f\") " pod="openshift-marketplace/redhat-marketplace-q8m72" Nov 25 14:22:00 crc kubenswrapper[4702]: I1125 14:22:00.212970 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/314da235-33f1-44f7-b9f0-3bc1af92879f-utilities\") pod \"redhat-marketplace-q8m72\" (UID: \"314da235-33f1-44f7-b9f0-3bc1af92879f\") " pod="openshift-marketplace/redhat-marketplace-q8m72" Nov 25 14:22:00 crc kubenswrapper[4702]: I1125 14:22:00.241309 4702 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnwhv\" (UniqueName: \"kubernetes.io/projected/314da235-33f1-44f7-b9f0-3bc1af92879f-kube-api-access-qnwhv\") pod \"redhat-marketplace-q8m72\" (UID: \"314da235-33f1-44f7-b9f0-3bc1af92879f\") " pod="openshift-marketplace/redhat-marketplace-q8m72" Nov 25 14:22:00 crc kubenswrapper[4702]: I1125 14:22:00.368414 4702 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q8m72" Nov 25 14:22:00 crc kubenswrapper[4702]: I1125 14:22:00.841431 4702 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q8m72"] Nov 25 14:22:01 crc kubenswrapper[4702]: I1125 14:22:01.701192 4702 generic.go:334] "Generic (PLEG): container finished" podID="314da235-33f1-44f7-b9f0-3bc1af92879f" containerID="52fb6c03f7c8a6a4203dbf4832bbd1b561af3ce4e84faf3b027a570347a7b0e5" exitCode=0 Nov 25 14:22:01 crc kubenswrapper[4702]: I1125 14:22:01.701272 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q8m72" event={"ID":"314da235-33f1-44f7-b9f0-3bc1af92879f","Type":"ContainerDied","Data":"52fb6c03f7c8a6a4203dbf4832bbd1b561af3ce4e84faf3b027a570347a7b0e5"} Nov 25 14:22:01 crc kubenswrapper[4702]: I1125 14:22:01.701316 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q8m72" event={"ID":"314da235-33f1-44f7-b9f0-3bc1af92879f","Type":"ContainerStarted","Data":"30f271beb4da4b9529a192df84ccdfeca14d4beb664e225f3c18f395b670a5e3"} Nov 25 14:22:03 crc kubenswrapper[4702]: I1125 14:22:03.719522 4702 generic.go:334] "Generic (PLEG): container finished" podID="314da235-33f1-44f7-b9f0-3bc1af92879f" containerID="f1af7e49383657503a98fe9448e48681a5aee67d17c032c9ba73183f2c27d5c5" exitCode=0 Nov 25 14:22:03 crc kubenswrapper[4702]: I1125 14:22:03.719639 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q8m72" event={"ID":"314da235-33f1-44f7-b9f0-3bc1af92879f","Type":"ContainerDied","Data":"f1af7e49383657503a98fe9448e48681a5aee67d17c032c9ba73183f2c27d5c5"} Nov 25 14:22:03 crc kubenswrapper[4702]: I1125 14:22:03.962624 4702 patch_prober.go:28] interesting pod/machine-config-daemon-qqr76 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 14:22:03 crc kubenswrapper[4702]: I1125 14:22:03.962946 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 14:22:04 crc kubenswrapper[4702]: I1125 14:22:04.727267 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q8m72" event={"ID":"314da235-33f1-44f7-b9f0-3bc1af92879f","Type":"ContainerStarted","Data":"a2bf66ebda9325b3b1029f61d7ee52b0454b8fde705a31ddf4bbe9e30df61748"} Nov 25 14:22:10 crc kubenswrapper[4702]: I1125 14:22:10.369307 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-q8m72" Nov 25 14:22:10 crc kubenswrapper[4702]: I1125 14:22:10.369824 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-q8m72" Nov 25 14:22:10 crc kubenswrapper[4702]: I1125 14:22:10.414312 4702 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-q8m72" Nov 25 14:22:10 crc kubenswrapper[4702]: I1125 14:22:10.451014 4702 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-q8m72" podStartSLOduration=8.03664513 podStartE2EDuration="10.450538837s" podCreationTimestamp="2025-11-25 14:22:00 +0000 UTC" firstStartedPulling="2025-11-25 14:22:01.707364679 +0000 UTC m=+2130.870015129" lastFinishedPulling="2025-11-25 14:22:04.121258386 +0000 UTC m=+2133.283908836" observedRunningTime="2025-11-25 14:22:04.759123298 +0000 UTC m=+2133.921773748" watchObservedRunningTime="2025-11-25 14:22:10.450538837 +0000 UTC m=+2139.613189287" Nov 25 14:22:10 crc kubenswrapper[4702]: I1125 14:22:10.822732 4702 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-q8m72" Nov 25 14:22:10 crc kubenswrapper[4702]: I1125 14:22:10.866852 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q8m72"] Nov 25 14:22:12 crc kubenswrapper[4702]: I1125 14:22:12.785730 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-q8m72" podUID="314da235-33f1-44f7-b9f0-3bc1af92879f" containerName="registry-server" containerID="cri-o://a2bf66ebda9325b3b1029f61d7ee52b0454b8fde705a31ddf4bbe9e30df61748" gracePeriod=2 Nov 25 14:22:13 crc kubenswrapper[4702]: I1125 14:22:13.246460 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q8m72" Nov 25 14:22:13 crc kubenswrapper[4702]: I1125 14:22:13.335286 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/314da235-33f1-44f7-b9f0-3bc1af92879f-utilities\") pod \"314da235-33f1-44f7-b9f0-3bc1af92879f\" (UID: \"314da235-33f1-44f7-b9f0-3bc1af92879f\") " Nov 25 14:22:13 crc kubenswrapper[4702]: I1125 14:22:13.335560 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/314da235-33f1-44f7-b9f0-3bc1af92879f-catalog-content\") pod \"314da235-33f1-44f7-b9f0-3bc1af92879f\" (UID: \"314da235-33f1-44f7-b9f0-3bc1af92879f\") " Nov 25 14:22:13 crc kubenswrapper[4702]: I1125 14:22:13.335687 4702 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnwhv\" (UniqueName: \"kubernetes.io/projected/314da235-33f1-44f7-b9f0-3bc1af92879f-kube-api-access-qnwhv\") pod \"314da235-33f1-44f7-b9f0-3bc1af92879f\" (UID: \"314da235-33f1-44f7-b9f0-3bc1af92879f\") " Nov 25 14:22:13 crc kubenswrapper[4702]: I1125 14:22:13.336481 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/314da235-33f1-44f7-b9f0-3bc1af92879f-utilities" (OuterVolumeSpecName: "utilities") pod "314da235-33f1-44f7-b9f0-3bc1af92879f" (UID: "314da235-33f1-44f7-b9f0-3bc1af92879f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:22:13 crc kubenswrapper[4702]: I1125 14:22:13.337557 4702 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/314da235-33f1-44f7-b9f0-3bc1af92879f-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 14:22:13 crc kubenswrapper[4702]: I1125 14:22:13.346227 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/314da235-33f1-44f7-b9f0-3bc1af92879f-kube-api-access-qnwhv" (OuterVolumeSpecName: "kube-api-access-qnwhv") pod "314da235-33f1-44f7-b9f0-3bc1af92879f" (UID: "314da235-33f1-44f7-b9f0-3bc1af92879f"). InnerVolumeSpecName "kube-api-access-qnwhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 14:22:13 crc kubenswrapper[4702]: I1125 14:22:13.365095 4702 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/314da235-33f1-44f7-b9f0-3bc1af92879f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "314da235-33f1-44f7-b9f0-3bc1af92879f" (UID: "314da235-33f1-44f7-b9f0-3bc1af92879f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 14:22:13 crc kubenswrapper[4702]: I1125 14:22:13.438554 4702 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/314da235-33f1-44f7-b9f0-3bc1af92879f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 14:22:13 crc kubenswrapper[4702]: I1125 14:22:13.438622 4702 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnwhv\" (UniqueName: \"kubernetes.io/projected/314da235-33f1-44f7-b9f0-3bc1af92879f-kube-api-access-qnwhv\") on node \"crc\" DevicePath \"\"" Nov 25 14:22:13 crc kubenswrapper[4702]: I1125 14:22:13.792211 4702 generic.go:334] "Generic (PLEG): container finished" podID="314da235-33f1-44f7-b9f0-3bc1af92879f" containerID="a2bf66ebda9325b3b1029f61d7ee52b0454b8fde705a31ddf4bbe9e30df61748" exitCode=0 Nov 25 14:22:13 crc kubenswrapper[4702]: I1125 14:22:13.792302 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q8m72" event={"ID":"314da235-33f1-44f7-b9f0-3bc1af92879f","Type":"ContainerDied","Data":"a2bf66ebda9325b3b1029f61d7ee52b0454b8fde705a31ddf4bbe9e30df61748"} Nov 25 14:22:13 crc kubenswrapper[4702]: I1125 14:22:13.792331 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q8m72" event={"ID":"314da235-33f1-44f7-b9f0-3bc1af92879f","Type":"ContainerDied","Data":"30f271beb4da4b9529a192df84ccdfeca14d4beb664e225f3c18f395b670a5e3"} Nov 25 14:22:13 crc kubenswrapper[4702]: I1125 14:22:13.792349 4702 scope.go:117] "RemoveContainer" containerID="a2bf66ebda9325b3b1029f61d7ee52b0454b8fde705a31ddf4bbe9e30df61748" Nov 25 14:22:13 crc kubenswrapper[4702]: I1125 14:22:13.792474 4702 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q8m72" Nov 25 14:22:13 crc kubenswrapper[4702]: I1125 14:22:13.825078 4702 scope.go:117] "RemoveContainer" containerID="f1af7e49383657503a98fe9448e48681a5aee67d17c032c9ba73183f2c27d5c5" Nov 25 14:22:13 crc kubenswrapper[4702]: I1125 14:22:13.829015 4702 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q8m72"] Nov 25 14:22:13 crc kubenswrapper[4702]: I1125 14:22:13.834280 4702 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-q8m72"] Nov 25 14:22:13 crc kubenswrapper[4702]: I1125 14:22:13.846284 4702 scope.go:117] "RemoveContainer" containerID="52fb6c03f7c8a6a4203dbf4832bbd1b561af3ce4e84faf3b027a570347a7b0e5" Nov 25 14:22:13 crc kubenswrapper[4702]: I1125 14:22:13.871953 4702 scope.go:117] "RemoveContainer" containerID="a2bf66ebda9325b3b1029f61d7ee52b0454b8fde705a31ddf4bbe9e30df61748" Nov 25 14:22:13 crc kubenswrapper[4702]: E1125 14:22:13.872318 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2bf66ebda9325b3b1029f61d7ee52b0454b8fde705a31ddf4bbe9e30df61748\": container with ID starting with a2bf66ebda9325b3b1029f61d7ee52b0454b8fde705a31ddf4bbe9e30df61748 not found: ID does not exist" containerID="a2bf66ebda9325b3b1029f61d7ee52b0454b8fde705a31ddf4bbe9e30df61748" Nov 25 14:22:13 crc kubenswrapper[4702]: I1125 14:22:13.872370 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2bf66ebda9325b3b1029f61d7ee52b0454b8fde705a31ddf4bbe9e30df61748"} err="failed to get container status \"a2bf66ebda9325b3b1029f61d7ee52b0454b8fde705a31ddf4bbe9e30df61748\": rpc error: code = NotFound desc = could not find container \"a2bf66ebda9325b3b1029f61d7ee52b0454b8fde705a31ddf4bbe9e30df61748\": container with ID starting with a2bf66ebda9325b3b1029f61d7ee52b0454b8fde705a31ddf4bbe9e30df61748 not found: ID does not exist" Nov 25 14:22:13 crc kubenswrapper[4702]: I1125 14:22:13.872392 4702 scope.go:117] "RemoveContainer" containerID="f1af7e49383657503a98fe9448e48681a5aee67d17c032c9ba73183f2c27d5c5" Nov 25 14:22:13 crc kubenswrapper[4702]: E1125 14:22:13.872685 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1af7e49383657503a98fe9448e48681a5aee67d17c032c9ba73183f2c27d5c5\": container with ID starting with f1af7e49383657503a98fe9448e48681a5aee67d17c032c9ba73183f2c27d5c5 not found: ID does not exist" containerID="f1af7e49383657503a98fe9448e48681a5aee67d17c032c9ba73183f2c27d5c5" Nov 25 14:22:13 crc kubenswrapper[4702]: I1125 14:22:13.872713 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1af7e49383657503a98fe9448e48681a5aee67d17c032c9ba73183f2c27d5c5"} err="failed to get container status \"f1af7e49383657503a98fe9448e48681a5aee67d17c032c9ba73183f2c27d5c5\": rpc error: code = NotFound desc = could not find container \"f1af7e49383657503a98fe9448e48681a5aee67d17c032c9ba73183f2c27d5c5\": container with ID starting with f1af7e49383657503a98fe9448e48681a5aee67d17c032c9ba73183f2c27d5c5 not found: ID does not exist" Nov 25 14:22:13 crc kubenswrapper[4702]: I1125 14:22:13.872747 4702 scope.go:117] "RemoveContainer" containerID="52fb6c03f7c8a6a4203dbf4832bbd1b561af3ce4e84faf3b027a570347a7b0e5" Nov 25 14:22:13 crc kubenswrapper[4702]: E1125 14:22:13.872934 4702 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52fb6c03f7c8a6a4203dbf4832bbd1b561af3ce4e84faf3b027a570347a7b0e5\": container with ID starting with 52fb6c03f7c8a6a4203dbf4832bbd1b561af3ce4e84faf3b027a570347a7b0e5 not found: ID does not exist" containerID="52fb6c03f7c8a6a4203dbf4832bbd1b561af3ce4e84faf3b027a570347a7b0e5" Nov 25 14:22:13 crc kubenswrapper[4702]: I1125 14:22:13.872973 4702 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52fb6c03f7c8a6a4203dbf4832bbd1b561af3ce4e84faf3b027a570347a7b0e5"} err="failed to get container status \"52fb6c03f7c8a6a4203dbf4832bbd1b561af3ce4e84faf3b027a570347a7b0e5\": rpc error: code = NotFound desc = could not find container \"52fb6c03f7c8a6a4203dbf4832bbd1b561af3ce4e84faf3b027a570347a7b0e5\": container with ID starting with 52fb6c03f7c8a6a4203dbf4832bbd1b561af3ce4e84faf3b027a570347a7b0e5 not found: ID does not exist" Nov 25 14:22:15 crc kubenswrapper[4702]: I1125 14:22:15.814590 4702 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="314da235-33f1-44f7-b9f0-3bc1af92879f" path="/var/lib/kubelet/pods/314da235-33f1-44f7-b9f0-3bc1af92879f/volumes" Nov 25 14:22:28 crc kubenswrapper[4702]: I1125 14:22:28.287804 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-k4rxf" podUID="058da3b1-d848-4742-9454-f980618f76bd" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Nov 25 14:22:33 crc kubenswrapper[4702]: I1125 14:22:33.943199 4702 scope.go:117] "RemoveContainer" containerID="343a84620bdc6ee504942e6a904d4bf9437e49094c13f6176cabb793a0524bbb" Nov 25 14:22:33 crc kubenswrapper[4702]: I1125 14:22:33.963215 4702 scope.go:117] "RemoveContainer" containerID="d89d654239cd9f7f93a9910cdfea090141ef0f2cb7223575974745f290431aa1" Nov 25 14:22:33 crc kubenswrapper[4702]: I1125 14:22:33.964097 4702 patch_prober.go:28] interesting pod/machine-config-daemon-qqr76 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 14:22:33 crc kubenswrapper[4702]: I1125 14:22:33.964260 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 14:22:34 crc kubenswrapper[4702]: I1125 14:22:34.024435 4702 scope.go:117] "RemoveContainer" containerID="0536a3dac6b639535a8e04cbcbdb471a210f9dd77cf0b1d4391a71227e1d1e17" Nov 25 14:23:03 crc kubenswrapper[4702]: I1125 14:23:03.962365 4702 patch_prober.go:28] interesting pod/machine-config-daemon-qqr76 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 14:23:03 crc kubenswrapper[4702]: I1125 14:23:03.962944 4702 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 14:23:03 crc kubenswrapper[4702]: I1125 14:23:03.963000 4702 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" Nov 25 14:23:03 crc kubenswrapper[4702]: I1125 14:23:03.963662 4702 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9f8fe70e21680c908554f1b846512eb5113ef4d904ef8a32b1e094b610045301"} pod="openshift-machine-config-operator/machine-config-daemon-qqr76" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 14:23:03 crc kubenswrapper[4702]: I1125 14:23:03.963727 4702 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" podUID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerName="machine-config-daemon" containerID="cri-o://9f8fe70e21680c908554f1b846512eb5113ef4d904ef8a32b1e094b610045301" gracePeriod=600 Nov 25 14:23:04 crc kubenswrapper[4702]: I1125 14:23:04.223163 4702 generic.go:334] "Generic (PLEG): container finished" podID="3c491818-b0bb-4d82-a031-96e2dfac8c27" containerID="9f8fe70e21680c908554f1b846512eb5113ef4d904ef8a32b1e094b610045301" exitCode=0 Nov 25 14:23:04 crc kubenswrapper[4702]: I1125 14:23:04.223228 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" event={"ID":"3c491818-b0bb-4d82-a031-96e2dfac8c27","Type":"ContainerDied","Data":"9f8fe70e21680c908554f1b846512eb5113ef4d904ef8a32b1e094b610045301"} Nov 25 14:23:04 crc kubenswrapper[4702]: I1125 14:23:04.223598 4702 scope.go:117] "RemoveContainer" containerID="32273717263d7fcc99e8193a439dd976da66869c55bb4633d9314cacb6ab92a1" Nov 25 14:23:05 crc kubenswrapper[4702]: I1125 14:23:05.231033 4702 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qqr76" event={"ID":"3c491818-b0bb-4d82-a031-96e2dfac8c27","Type":"ContainerStarted","Data":"1aec4d174bdd252a40ef013bcd9bc99e4354bd55bd88fe7526976b146fd995b5"} var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515111335750024446 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015111335750017363 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015111331221016473 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015111331221015443 5ustar corecore